var/home/core/zuul-output/0000755000175000017500000000000015112302370014517 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112307253015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004613757315112307244017714 0ustar rootrootNov 28 11:22:01 crc systemd[1]: Starting Kubernetes Kubelet... Nov 28 11:22:01 crc restorecon[4742]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:01 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 11:22:02 crc restorecon[4742]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 28 11:22:02 crc kubenswrapper[4862]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 11:22:02 crc kubenswrapper[4862]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 28 11:22:02 crc kubenswrapper[4862]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 11:22:02 crc kubenswrapper[4862]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 11:22:02 crc kubenswrapper[4862]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 28 11:22:02 crc kubenswrapper[4862]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.603636 4862 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606574 4862 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606592 4862 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606597 4862 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606602 4862 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606605 4862 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606609 4862 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606613 4862 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606617 4862 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606622 4862 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606625 4862 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606629 4862 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606633 4862 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606638 4862 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606643 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606647 4862 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606653 4862 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606657 4862 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606662 4862 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606667 4862 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606671 4862 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606676 4862 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606680 4862 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606683 4862 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606687 4862 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606690 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606694 4862 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606697 4862 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606701 4862 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606705 4862 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606708 4862 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606712 4862 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606715 4862 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606718 4862 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606723 4862 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606727 4862 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606730 4862 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606734 4862 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606737 4862 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606741 4862 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606745 4862 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606750 4862 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606754 4862 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606758 4862 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606761 4862 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606765 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606769 4862 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606772 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606777 4862 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606781 4862 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606785 4862 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606788 4862 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606792 4862 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606805 4862 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606809 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606813 4862 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606817 4862 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606821 4862 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606824 4862 feature_gate.go:330] unrecognized feature gate: Example Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606827 4862 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606831 4862 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606834 4862 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606838 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606841 4862 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606846 4862 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606851 4862 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606855 4862 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606859 4862 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606863 4862 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606866 4862 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606870 4862 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.606873 4862 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.606952 4862 flags.go:64] FLAG: --address="0.0.0.0" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.606960 4862 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.606968 4862 flags.go:64] FLAG: --anonymous-auth="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.606973 4862 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.606978 4862 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.606982 4862 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.606988 4862 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.606993 4862 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.606997 4862 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607001 4862 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607006 4862 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607010 4862 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607014 4862 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607021 4862 flags.go:64] FLAG: --cgroup-root="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607026 4862 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607031 4862 flags.go:64] FLAG: --client-ca-file="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607035 4862 flags.go:64] FLAG: --cloud-config="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607039 4862 flags.go:64] FLAG: --cloud-provider="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607043 4862 flags.go:64] FLAG: --cluster-dns="[]" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607048 4862 flags.go:64] FLAG: --cluster-domain="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607052 4862 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607057 4862 flags.go:64] FLAG: --config-dir="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607061 4862 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607066 4862 flags.go:64] FLAG: --container-log-max-files="5" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607071 4862 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607076 4862 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607081 4862 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607085 4862 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607104 4862 flags.go:64] FLAG: --contention-profiling="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607108 4862 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607113 4862 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607117 4862 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607121 4862 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607127 4862 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607131 4862 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607135 4862 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607139 4862 flags.go:64] FLAG: --enable-load-reader="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607143 4862 flags.go:64] FLAG: --enable-server="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607147 4862 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607153 4862 flags.go:64] FLAG: --event-burst="100" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607158 4862 flags.go:64] FLAG: --event-qps="50" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607163 4862 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607167 4862 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607171 4862 flags.go:64] FLAG: --eviction-hard="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607176 4862 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607181 4862 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607185 4862 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607189 4862 flags.go:64] FLAG: --eviction-soft="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607193 4862 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607198 4862 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607202 4862 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607206 4862 flags.go:64] FLAG: --experimental-mounter-path="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607210 4862 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607214 4862 flags.go:64] FLAG: --fail-swap-on="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607218 4862 flags.go:64] FLAG: --feature-gates="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607223 4862 flags.go:64] FLAG: --file-check-frequency="20s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607227 4862 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607232 4862 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607236 4862 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607241 4862 flags.go:64] FLAG: --healthz-port="10248" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607245 4862 flags.go:64] FLAG: --help="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607249 4862 flags.go:64] FLAG: --hostname-override="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607253 4862 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607257 4862 flags.go:64] FLAG: --http-check-frequency="20s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607261 4862 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607265 4862 flags.go:64] FLAG: --image-credential-provider-config="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607269 4862 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607273 4862 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607277 4862 flags.go:64] FLAG: --image-service-endpoint="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607281 4862 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607285 4862 flags.go:64] FLAG: --kube-api-burst="100" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607289 4862 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607293 4862 flags.go:64] FLAG: --kube-api-qps="50" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607297 4862 flags.go:64] FLAG: --kube-reserved="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607301 4862 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607305 4862 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607309 4862 flags.go:64] FLAG: --kubelet-cgroups="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607314 4862 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607318 4862 flags.go:64] FLAG: --lock-file="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607323 4862 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607327 4862 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607331 4862 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607337 4862 flags.go:64] FLAG: --log-json-split-stream="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607341 4862 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607345 4862 flags.go:64] FLAG: --log-text-split-stream="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607349 4862 flags.go:64] FLAG: --logging-format="text" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607353 4862 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607357 4862 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607362 4862 flags.go:64] FLAG: --manifest-url="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607365 4862 flags.go:64] FLAG: --manifest-url-header="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607371 4862 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607375 4862 flags.go:64] FLAG: --max-open-files="1000000" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607380 4862 flags.go:64] FLAG: --max-pods="110" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607384 4862 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607388 4862 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607392 4862 flags.go:64] FLAG: --memory-manager-policy="None" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607396 4862 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607400 4862 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607404 4862 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607408 4862 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607417 4862 flags.go:64] FLAG: --node-status-max-images="50" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607421 4862 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607425 4862 flags.go:64] FLAG: --oom-score-adj="-999" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607450 4862 flags.go:64] FLAG: --pod-cidr="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607454 4862 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607464 4862 flags.go:64] FLAG: --pod-manifest-path="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607468 4862 flags.go:64] FLAG: --pod-max-pids="-1" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607473 4862 flags.go:64] FLAG: --pods-per-core="0" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607477 4862 flags.go:64] FLAG: --port="10250" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607481 4862 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607486 4862 flags.go:64] FLAG: --provider-id="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607490 4862 flags.go:64] FLAG: --qos-reserved="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607494 4862 flags.go:64] FLAG: --read-only-port="10255" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607498 4862 flags.go:64] FLAG: --register-node="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607503 4862 flags.go:64] FLAG: --register-schedulable="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607507 4862 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607514 4862 flags.go:64] FLAG: --registry-burst="10" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607518 4862 flags.go:64] FLAG: --registry-qps="5" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607522 4862 flags.go:64] FLAG: --reserved-cpus="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607526 4862 flags.go:64] FLAG: --reserved-memory="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607531 4862 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607536 4862 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607541 4862 flags.go:64] FLAG: --rotate-certificates="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607545 4862 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607549 4862 flags.go:64] FLAG: --runonce="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607553 4862 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607558 4862 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607562 4862 flags.go:64] FLAG: --seccomp-default="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607566 4862 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607570 4862 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607574 4862 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607578 4862 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607582 4862 flags.go:64] FLAG: --storage-driver-password="root" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607586 4862 flags.go:64] FLAG: --storage-driver-secure="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607590 4862 flags.go:64] FLAG: --storage-driver-table="stats" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607594 4862 flags.go:64] FLAG: --storage-driver-user="root" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607598 4862 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607602 4862 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607606 4862 flags.go:64] FLAG: --system-cgroups="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607610 4862 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607617 4862 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607621 4862 flags.go:64] FLAG: --tls-cert-file="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607630 4862 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607635 4862 flags.go:64] FLAG: --tls-min-version="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607639 4862 flags.go:64] FLAG: --tls-private-key-file="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607643 4862 flags.go:64] FLAG: --topology-manager-policy="none" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607647 4862 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607651 4862 flags.go:64] FLAG: --topology-manager-scope="container" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607655 4862 flags.go:64] FLAG: --v="2" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607660 4862 flags.go:64] FLAG: --version="false" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607666 4862 flags.go:64] FLAG: --vmodule="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607671 4862 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.607675 4862 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607782 4862 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607787 4862 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607791 4862 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607794 4862 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607799 4862 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607803 4862 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607807 4862 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607810 4862 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607814 4862 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607818 4862 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607822 4862 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607827 4862 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607831 4862 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607834 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607838 4862 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607842 4862 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607846 4862 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607849 4862 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607854 4862 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607858 4862 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607863 4862 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607869 4862 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607873 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607877 4862 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607881 4862 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607884 4862 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607888 4862 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607892 4862 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607897 4862 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607901 4862 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607906 4862 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607909 4862 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607913 4862 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607917 4862 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607921 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607925 4862 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607928 4862 feature_gate.go:330] unrecognized feature gate: Example Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607932 4862 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607936 4862 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607940 4862 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607946 4862 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607950 4862 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607954 4862 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607958 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607965 4862 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607968 4862 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607972 4862 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607976 4862 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607980 4862 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607984 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607987 4862 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607991 4862 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.607995 4862 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608000 4862 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608004 4862 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608008 4862 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608011 4862 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608015 4862 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608019 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608023 4862 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608027 4862 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608030 4862 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608034 4862 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608038 4862 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608041 4862 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608045 4862 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608049 4862 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608052 4862 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608056 4862 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608060 4862 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.608065 4862 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.608230 4862 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.619853 4862 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.619929 4862 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620080 4862 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620124 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620136 4862 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620147 4862 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620162 4862 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620177 4862 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620187 4862 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620196 4862 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620205 4862 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620216 4862 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620227 4862 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620237 4862 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620247 4862 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620257 4862 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620267 4862 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620276 4862 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620284 4862 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620292 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620299 4862 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620307 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620315 4862 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620324 4862 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620332 4862 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620340 4862 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620347 4862 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620355 4862 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620364 4862 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620371 4862 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620379 4862 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620387 4862 feature_gate.go:330] unrecognized feature gate: Example Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620397 4862 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620406 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620414 4862 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620421 4862 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620429 4862 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620436 4862 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620444 4862 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620452 4862 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620460 4862 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620468 4862 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620476 4862 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620484 4862 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620492 4862 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620499 4862 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620507 4862 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620515 4862 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620523 4862 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620531 4862 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620539 4862 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620546 4862 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620554 4862 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620562 4862 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620570 4862 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620580 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620588 4862 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620595 4862 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620603 4862 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620610 4862 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620619 4862 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620629 4862 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620639 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620647 4862 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620655 4862 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620664 4862 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620672 4862 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620679 4862 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620688 4862 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620696 4862 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620704 4862 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620712 4862 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620720 4862 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.620733 4862 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620957 4862 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620969 4862 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620980 4862 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.620991 4862 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621001 4862 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621010 4862 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621018 4862 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621027 4862 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621036 4862 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621045 4862 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621054 4862 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621062 4862 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621070 4862 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621078 4862 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621087 4862 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621121 4862 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621129 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621137 4862 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621144 4862 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621153 4862 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621160 4862 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621168 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621176 4862 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621184 4862 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621191 4862 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621199 4862 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621206 4862 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621214 4862 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621222 4862 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621232 4862 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621243 4862 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621252 4862 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621261 4862 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621268 4862 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621277 4862 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621285 4862 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621295 4862 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621306 4862 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621315 4862 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621323 4862 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621332 4862 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621340 4862 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621349 4862 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621357 4862 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621364 4862 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621372 4862 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621380 4862 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621388 4862 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621396 4862 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621403 4862 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621411 4862 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621420 4862 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621428 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621436 4862 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621444 4862 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621451 4862 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621459 4862 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621466 4862 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621476 4862 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621484 4862 feature_gate.go:330] unrecognized feature gate: Example Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621492 4862 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621499 4862 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621507 4862 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621515 4862 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621522 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621530 4862 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.621639 4862 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.623978 4862 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.624030 4862 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.624041 4862 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.624054 4862 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.624071 4862 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.624949 4862 server.go:940] "Client rotation is on, will bootstrap in background" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.630082 4862 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.630262 4862 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.631168 4862 server.go:997] "Starting client certificate rotation" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.631221 4862 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.631424 4862 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-06 09:40:03.046977255 +0000 UTC Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.631588 4862 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 190h18m0.415395726s for next certificate rotation Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.640762 4862 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.643928 4862 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.655124 4862 log.go:25] "Validated CRI v1 runtime API" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.715595 4862 log.go:25] "Validated CRI v1 image API" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.718572 4862 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.722343 4862 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-28-11-17-03-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.722400 4862 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.750364 4862 manager.go:217] Machine: {Timestamp:2025-11-28 11:22:02.746031759 +0000 UTC m=+0.278545710 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:b1809799-f6d1-48c5-b4d0-d4f1bd5f5602 BootID:0c620f82-404c-4d65-9ba0-65e90c64ab19 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f1:de:0b Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f1:de:0b Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:95:21:fa Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e0:32:37 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:82:76:e7 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d2:d8:9d Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:9a:27:bc Speed:-1 Mtu:1496} {Name:eth10 MacAddress:7a:af:11:14:75:69 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:1a:10:cd:cf:14:64 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.750682 4862 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.751042 4862 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.751679 4862 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.751891 4862 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.751938 4862 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.752217 4862 topology_manager.go:138] "Creating topology manager with none policy" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.752231 4862 container_manager_linux.go:303] "Creating device plugin manager" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.752466 4862 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.752506 4862 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.752800 4862 state_mem.go:36] "Initialized new in-memory state store" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.752899 4862 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.754277 4862 kubelet.go:418] "Attempting to sync node with API server" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.754312 4862 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.754349 4862 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.754366 4862 kubelet.go:324] "Adding apiserver pod source" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.754380 4862 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.756703 4862 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.757068 4862 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.757838 4862 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.757927 4862 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.757929 4862 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.758025 4862 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.758516 4862 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759287 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759316 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759325 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759334 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759348 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759358 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759370 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759389 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759407 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759416 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759429 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759462 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.759901 4862 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.760607 4862 server.go:1280] "Started kubelet" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.760822 4862 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.761075 4862 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.761630 4862 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.762223 4862 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 28 11:22:02 crc systemd[1]: Started Kubernetes Kubelet. Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.764939 4862 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.765004 4862 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.765065 4862 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 07:59:46.045667141 +0000 UTC Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.765175 4862 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 404h37m43.280497759s for next certificate rotation Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.765289 4862 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.765412 4862 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.765437 4862 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.765924 4862 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.765454 4862 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187c27ceb4177ac8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 11:22:02.760575688 +0000 UTC m=+0.293089619,LastTimestamp:2025-11-28 11:22:02.760575688 +0000 UTC m=+0.293089619,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.767549 4862 factory.go:153] Registering CRI-O factory Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.767600 4862 factory.go:221] Registration of the crio container factory successfully Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.767549 4862 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.767668 4862 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.767731 4862 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.767841 4862 factory.go:55] Registering systemd factory Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.767860 4862 factory.go:221] Registration of the systemd container factory successfully Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.767906 4862 factory.go:103] Registering Raw factory Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.767941 4862 manager.go:1196] Started watching for new ooms in manager Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.768193 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.769263 4862 server.go:460] "Adding debug handlers to kubelet server" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.769770 4862 manager.go:319] Starting recovery of all containers Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.786955 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.787439 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.787697 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.787888 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.788164 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.788390 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.788582 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.788773 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.788960 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.789183 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.789401 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.789595 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.789773 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.789961 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.790165 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.790367 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.790553 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.790748 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.790983 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.791203 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.791414 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.791629 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.791810 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.791997 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.792245 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.792471 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.792694 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.792888 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.793068 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.793363 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.793555 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.793772 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.793955 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.794180 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.794350 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.794515 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.794713 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.794885 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.795055 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.795288 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.795529 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.795736 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.795921 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.796143 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.796338 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.796528 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.796717 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.796926 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.797159 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.797398 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.797594 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.797790 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798016 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798260 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798431 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798602 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798812 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798854 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798871 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798886 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798900 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798914 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798926 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798942 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798955 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798967 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798979 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.798995 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799010 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799024 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799038 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799052 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799065 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799080 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799115 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799129 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799143 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799157 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799170 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799183 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799196 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799209 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799221 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799239 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799255 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799270 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799283 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799299 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799313 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799326 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799342 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799356 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799369 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799383 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799396 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799409 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799422 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799435 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799477 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799498 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799517 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799536 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799552 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799571 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799644 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799662 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799701 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799716 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799735 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799753 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799767 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799783 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799800 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799823 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799838 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799853 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799866 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799879 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799893 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799905 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799920 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799933 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799946 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799960 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799973 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799986 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.799999 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800013 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800027 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800040 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800055 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800068 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800082 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800129 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800143 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800157 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800171 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800185 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800199 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800214 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800228 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800244 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800258 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800273 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800290 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800306 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800322 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800341 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800357 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800372 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.800388 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801173 4862 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801210 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801229 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801286 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801304 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801320 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801335 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801350 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801367 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801384 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801398 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801412 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801429 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801443 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801458 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801471 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801485 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801501 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801516 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801530 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801544 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801563 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801578 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801593 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801608 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801623 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801638 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801652 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801670 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801687 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801705 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801722 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801737 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801752 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801767 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801782 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801797 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801812 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801826 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801841 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801858 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801873 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801889 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801903 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801918 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801932 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801946 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801961 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801975 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.801991 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.802006 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.802021 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.802036 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.802053 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.802068 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.802083 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.802117 4862 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.802131 4862 reconstruct.go:97] "Volume reconstruction finished" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.802141 4862 reconciler.go:26] "Reconciler: start to sync state" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.809352 4862 manager.go:324] Recovery completed Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.828067 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.830694 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.830763 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.830779 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.831694 4862 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.831718 4862 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.831743 4862 state_mem.go:36] "Initialized new in-memory state store" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.834424 4862 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.837318 4862 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.837362 4862 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.837395 4862 kubelet.go:2335] "Starting kubelet main sync loop" Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.837445 4862 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 28 11:22:02 crc kubenswrapper[4862]: W1128 11:22:02.843227 4862 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.843341 4862 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.843749 4862 policy_none.go:49] "None policy: Start" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.848162 4862 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.848201 4862 state_mem.go:35] "Initializing new in-memory state store" Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.865787 4862 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.906270 4862 manager.go:334] "Starting Device Plugin manager" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.906365 4862 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.906387 4862 server.go:79] "Starting device plugin registration server" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.907429 4862 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.907480 4862 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.909294 4862 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.909427 4862 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.909446 4862 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.921911 4862 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.938157 4862 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.938287 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.939622 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.939657 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.939666 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.939781 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.940329 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.940446 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.940567 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.940627 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.940647 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.940896 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.941273 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.941362 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.942003 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.942049 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.942068 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.943562 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.943590 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.943604 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.943788 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.943806 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.943816 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.943972 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.944495 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.944580 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.945984 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.946010 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.946021 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.946168 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.946716 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.946748 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947079 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947123 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947133 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947612 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947632 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947644 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947800 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947816 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947826 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947969 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.947999 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.949583 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.949617 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:02 crc kubenswrapper[4862]: I1128 11:22:02.949656 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:02 crc kubenswrapper[4862]: E1128 11:22:02.969920 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004316 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004409 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004458 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004496 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004534 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004571 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004602 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004633 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004667 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004696 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004728 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004759 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004788 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004820 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.004884 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.008473 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.010038 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.010084 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.010113 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.010144 4862 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 11:22:03 crc kubenswrapper[4862]: E1128 11:22:03.010855 4862 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.105814 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.105909 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.105971 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106021 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106062 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106143 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106191 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106236 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106279 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106314 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106305 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106372 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106430 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106463 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106345 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106711 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106515 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106537 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106549 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106551 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106545 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106818 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106492 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.106851 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.107006 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.107064 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.107076 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.107149 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.107176 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.107233 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.211004 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.213201 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.213275 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.213295 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.213340 4862 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 11:22:03 crc kubenswrapper[4862]: E1128 11:22:03.214173 4862 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.277242 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.287991 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.310145 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: W1128 11:22:03.313819 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2ba999fdab8f6f2d8d82398f67e7e8ce2bb83fdfdc183a7e2647e683da03cbef WatchSource:0}: Error finding container 2ba999fdab8f6f2d8d82398f67e7e8ce2bb83fdfdc183a7e2647e683da03cbef: Status 404 returned error can't find the container with id 2ba999fdab8f6f2d8d82398f67e7e8ce2bb83fdfdc183a7e2647e683da03cbef Nov 28 11:22:03 crc kubenswrapper[4862]: W1128 11:22:03.318707 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-916d57f4079ffdd90ef21493c8329a2004dc8f8ab3a0af4f4e4d1915f3057316 WatchSource:0}: Error finding container 916d57f4079ffdd90ef21493c8329a2004dc8f8ab3a0af4f4e4d1915f3057316: Status 404 returned error can't find the container with id 916d57f4079ffdd90ef21493c8329a2004dc8f8ab3a0af4f4e4d1915f3057316 Nov 28 11:22:03 crc kubenswrapper[4862]: W1128 11:22:03.331579 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-01c6e8299561ef51099b5292906ba16292759222a487aea52cfdf269ae33cd82 WatchSource:0}: Error finding container 01c6e8299561ef51099b5292906ba16292759222a487aea52cfdf269ae33cd82: Status 404 returned error can't find the container with id 01c6e8299561ef51099b5292906ba16292759222a487aea52cfdf269ae33cd82 Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.331730 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.340996 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:03 crc kubenswrapper[4862]: E1128 11:22:03.371313 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Nov 28 11:22:03 crc kubenswrapper[4862]: W1128 11:22:03.376037 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-e11bd236840f21a99f66d4cecf3be476c8593102e5157cc880641dfaa52520af WatchSource:0}: Error finding container e11bd236840f21a99f66d4cecf3be476c8593102e5157cc880641dfaa52520af: Status 404 returned error can't find the container with id e11bd236840f21a99f66d4cecf3be476c8593102e5157cc880641dfaa52520af Nov 28 11:22:03 crc kubenswrapper[4862]: W1128 11:22:03.379814 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ec3c14f1bbe5d02b363a2e1bc10eeabcf1196b689c9fd43e1d1ecaf765bc77a8 WatchSource:0}: Error finding container ec3c14f1bbe5d02b363a2e1bc10eeabcf1196b689c9fd43e1d1ecaf765bc77a8: Status 404 returned error can't find the container with id ec3c14f1bbe5d02b363a2e1bc10eeabcf1196b689c9fd43e1d1ecaf765bc77a8 Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.614385 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.616229 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.616270 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.616284 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.616316 4862 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 11:22:03 crc kubenswrapper[4862]: E1128 11:22:03.616777 4862 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Nov 28 11:22:03 crc kubenswrapper[4862]: W1128 11:22:03.637550 4862 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:03 crc kubenswrapper[4862]: E1128 11:22:03.637651 4862 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.762589 4862 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.851195 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.851379 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2ba999fdab8f6f2d8d82398f67e7e8ce2bb83fdfdc183a7e2647e683da03cbef"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.853118 4862 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8" exitCode=0 Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.853162 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.853215 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"916d57f4079ffdd90ef21493c8329a2004dc8f8ab3a0af4f4e4d1915f3057316"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.853403 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.855730 4862 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87" exitCode=0 Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.855807 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.855836 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ec3c14f1bbe5d02b363a2e1bc10eeabcf1196b689c9fd43e1d1ecaf765bc77a8"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.855930 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.856025 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.856119 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.856134 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.856771 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.856796 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.856803 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.858245 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.858678 4862 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e" exitCode=0 Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.858733 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.858869 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e11bd236840f21a99f66d4cecf3be476c8593102e5157cc880641dfaa52520af"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.858785 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.858950 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.858965 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.859192 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.861121 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.861172 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.861186 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.861581 4862 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a01fb725ec46ee8b2588256779d17793f286445d63eb0e0a2acef74beac25678" exitCode=0 Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.861636 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a01fb725ec46ee8b2588256779d17793f286445d63eb0e0a2acef74beac25678"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.861675 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"01c6e8299561ef51099b5292906ba16292759222a487aea52cfdf269ae33cd82"} Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.861765 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.862512 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.862543 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:03 crc kubenswrapper[4862]: I1128 11:22:03.862556 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:03 crc kubenswrapper[4862]: W1128 11:22:03.995054 4862 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:03 crc kubenswrapper[4862]: E1128 11:22:03.995277 4862 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 28 11:22:04 crc kubenswrapper[4862]: E1128 11:22:04.172744 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Nov 28 11:22:04 crc kubenswrapper[4862]: W1128 11:22:04.210076 4862 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:04 crc kubenswrapper[4862]: E1128 11:22:04.210191 4862 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 28 11:22:04 crc kubenswrapper[4862]: W1128 11:22:04.395317 4862 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 28 11:22:04 crc kubenswrapper[4862]: E1128 11:22:04.395552 4862 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.417599 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.419247 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.419289 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.419299 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.419324 4862 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.867890 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d2f59dd4bb9483584434a420817fc9b55580712d43b6e8267e286e1b6e3acb83"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.868024 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.869919 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.869973 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.869991 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.871883 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.871903 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.871916 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.871972 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.873284 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.873327 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.873349 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.876375 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"87d5fb88089c845017e6226c86fd1a41fc3622ad32b9ffec123ebb3649be4c58"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.876419 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"069da5c66373778aec270d3597c665d396a164a6c07f258553405d7b269f3f3d"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.876437 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3ca7d0113725fb7adcdea09bd4592580b9652ce2bb40e095484ec0c458203eed"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.876552 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.877696 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.877733 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.877750 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.880677 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.880713 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.880735 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.880754 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.882789 4862 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b" exitCode=0 Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.882827 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b"} Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.882971 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.883960 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.883990 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:04 crc kubenswrapper[4862]: I1128 11:22:04.884000 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.162181 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.171620 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.873857 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.894857 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82"} Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.895026 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.896908 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.896982 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.897007 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.899273 4862 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732" exitCode=0 Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.899349 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732"} Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.899502 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.899540 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.900980 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.901016 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.901052 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.901069 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.901027 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:05 crc kubenswrapper[4862]: I1128 11:22:05.901198 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.173702 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.173920 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.175337 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.175408 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.175427 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.618444 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.638673 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.909132 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7"} Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.909214 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c"} Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.909231 4862 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.909245 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200"} Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.909290 4862 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.909369 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.909308 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.910998 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.911062 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.911087 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.911370 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.911418 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:06 crc kubenswrapper[4862]: I1128 11:22:06.911433 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.920309 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c"} Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.920385 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102"} Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.920421 4862 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.920467 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.920506 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.922387 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.922448 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.922485 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.922501 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.922450 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:07 crc kubenswrapper[4862]: I1128 11:22:07.922570 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.015078 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.332536 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.332783 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.334550 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.334626 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.334656 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.714063 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.906393 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.923476 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.923568 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.923729 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.925295 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.925346 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.925366 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.925371 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.925401 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.925417 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.925727 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.925772 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:08 crc kubenswrapper[4862]: I1128 11:22:08.925790 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:09 crc kubenswrapper[4862]: I1128 11:22:09.925878 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:09 crc kubenswrapper[4862]: I1128 11:22:09.927251 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:09 crc kubenswrapper[4862]: I1128 11:22:09.927333 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:09 crc kubenswrapper[4862]: I1128 11:22:09.927355 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:11 crc kubenswrapper[4862]: I1128 11:22:11.714508 4862 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 11:22:11 crc kubenswrapper[4862]: I1128 11:22:11.714626 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 11:22:12 crc kubenswrapper[4862]: E1128 11:22:12.922971 4862 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 11:22:13 crc kubenswrapper[4862]: I1128 11:22:13.526440 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 28 11:22:13 crc kubenswrapper[4862]: I1128 11:22:13.526651 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:13 crc kubenswrapper[4862]: I1128 11:22:13.527865 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:13 crc kubenswrapper[4862]: I1128 11:22:13.527895 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:13 crc kubenswrapper[4862]: I1128 11:22:13.527907 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:14 crc kubenswrapper[4862]: E1128 11:22:14.420814 4862 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Nov 28 11:22:14 crc kubenswrapper[4862]: I1128 11:22:14.763581 4862 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 28 11:22:15 crc kubenswrapper[4862]: E1128 11:22:15.161699 4862 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187c27ceb4177ac8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 11:22:02.760575688 +0000 UTC m=+0.293089619,LastTimestamp:2025-11-28 11:22:02.760575688 +0000 UTC m=+0.293089619,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 11:22:15 crc kubenswrapper[4862]: E1128 11:22:15.775130 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Nov 28 11:22:15 crc kubenswrapper[4862]: W1128 11:22:15.883934 4862 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 28 11:22:15 crc kubenswrapper[4862]: I1128 11:22:15.884763 4862 trace.go:236] Trace[336364014]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 11:22:05.879) (total time: 10005ms): Nov 28 11:22:15 crc kubenswrapper[4862]: Trace[336364014]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10004ms (11:22:15.883) Nov 28 11:22:15 crc kubenswrapper[4862]: Trace[336364014]: [10.005621789s] [10.005621789s] END Nov 28 11:22:15 crc kubenswrapper[4862]: E1128 11:22:15.884806 4862 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.021173 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.023010 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.023085 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.023116 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.023158 4862 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.279645 4862 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.279729 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.289382 4862 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.289459 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.649671 4862 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]log ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]etcd ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/generic-apiserver-start-informers ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/priority-and-fairness-filter ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-apiextensions-informers ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-apiextensions-controllers ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/crd-informer-synced ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-system-namespaces-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 28 11:22:16 crc kubenswrapper[4862]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 28 11:22:16 crc kubenswrapper[4862]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/bootstrap-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/start-kube-aggregator-informers ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/apiservice-registration-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/apiservice-discovery-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]autoregister-completion ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/apiservice-openapi-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 28 11:22:16 crc kubenswrapper[4862]: livez check failed Nov 28 11:22:16 crc kubenswrapper[4862]: I1128 11:22:16.649817 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:22:18 crc kubenswrapper[4862]: I1128 11:22:18.337887 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:18 crc kubenswrapper[4862]: I1128 11:22:18.338197 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:18 crc kubenswrapper[4862]: I1128 11:22:18.339955 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:18 crc kubenswrapper[4862]: I1128 11:22:18.339994 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:18 crc kubenswrapper[4862]: I1128 11:22:18.340007 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.270759 4862 trace.go:236] Trace[1380140301]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 11:22:06.421) (total time: 14849ms): Nov 28 11:22:21 crc kubenswrapper[4862]: Trace[1380140301]: ---"Objects listed" error: 14849ms (11:22:21.270) Nov 28 11:22:21 crc kubenswrapper[4862]: Trace[1380140301]: [14.849506506s] [14.849506506s] END Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.270823 4862 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.271571 4862 trace.go:236] Trace[31378674]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 11:22:06.358) (total time: 14913ms): Nov 28 11:22:21 crc kubenswrapper[4862]: Trace[31378674]: ---"Objects listed" error: 14913ms (11:22:21.271) Nov 28 11:22:21 crc kubenswrapper[4862]: Trace[31378674]: [14.913099217s] [14.913099217s] END Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.271598 4862 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.271724 4862 trace.go:236] Trace[753928389]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 11:22:06.760) (total time: 14510ms): Nov 28 11:22:21 crc kubenswrapper[4862]: Trace[753928389]: ---"Objects listed" error: 14510ms (11:22:21.271) Nov 28 11:22:21 crc kubenswrapper[4862]: Trace[753928389]: [14.510669512s] [14.510669512s] END Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.271765 4862 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.274759 4862 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.275537 4862 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.321866 4862 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42558->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.321966 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42558->192.168.126.11:17697: read: connection reset by peer" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.336361 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.348202 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.647732 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.653299 4862 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.653378 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.656424 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.764349 4862 apiserver.go:52] "Watching apiserver" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.767654 4862 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.768040 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.768510 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.768510 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.768617 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.768700 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.768868 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.768938 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.769155 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.769209 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.769502 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.773653 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.774016 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.774040 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.776575 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.776652 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.783728 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.786816 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.790145 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.795937 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.812085 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.828033 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.842878 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.857330 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.867534 4862 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880105 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880153 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880177 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880198 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880221 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880249 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880273 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880329 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880354 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880413 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880439 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880461 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880485 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880511 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880531 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880554 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880576 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880602 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880629 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880651 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880675 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880894 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880933 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880582 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881385 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880656 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880751 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880782 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880708 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880973 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880992 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881200 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881451 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881538 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881565 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.881640 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:22:22.381581181 +0000 UTC m=+19.914095102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881728 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881749 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.880952 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881794 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881813 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881833 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881834 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881890 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881918 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881934 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882105 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882128 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882251 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882243 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882266 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882304 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.881956 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882342 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882364 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882410 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882429 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882446 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882466 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882488 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882506 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882527 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882553 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882598 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882619 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882638 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882656 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882679 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882698 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882723 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882744 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882763 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882782 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882364 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882803 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882825 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882850 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882869 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882887 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882904 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882919 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882940 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882955 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882974 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882991 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883008 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883023 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883040 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883059 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883076 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883405 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883430 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883449 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883464 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883483 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883502 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883521 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883537 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883553 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883569 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883585 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883600 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883618 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883640 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883660 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883679 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883698 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883715 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883734 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883750 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883767 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883782 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883798 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883813 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883828 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883851 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883874 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883890 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883906 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883923 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883938 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883956 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883974 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883993 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884014 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884032 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884047 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884064 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882552 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884113 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882557 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882641 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882681 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882776 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882786 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882827 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.882847 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883045 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883079 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883257 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883326 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883599 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883610 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883720 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883814 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883854 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.883946 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884054 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884086 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884449 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884084 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884600 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884619 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884635 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884632 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884654 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884673 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884690 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884711 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884727 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884742 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884758 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884774 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884794 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884811 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884828 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884844 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884849 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884862 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884883 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884902 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884919 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884935 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884953 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884969 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.884987 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885002 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885019 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885035 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885052 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885067 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885103 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885119 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885157 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885175 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885192 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885207 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885223 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885242 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885258 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885275 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885296 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885314 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885367 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885385 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885403 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885420 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885437 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885718 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885738 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885755 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885773 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885793 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885812 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885833 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885864 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885884 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885902 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885924 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885943 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885961 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885980 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886000 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886022 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886041 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886061 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886082 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886115 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886133 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886151 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886173 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886191 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886210 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886229 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886249 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886266 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886285 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886311 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886332 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886353 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886371 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886395 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886416 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886435 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886467 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886486 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886504 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886521 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886545 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886562 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886580 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886599 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886615 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886657 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886681 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886702 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886720 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886745 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886765 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886786 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886813 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886839 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886867 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886890 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886911 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886937 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886965 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887026 4862 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887040 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887053 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887064 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887074 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887084 4862 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887114 4862 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887124 4862 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887135 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887145 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887154 4862 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887164 4862 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887174 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887184 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887193 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887206 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887216 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887225 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887235 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887246 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887258 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887272 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887285 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887299 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887311 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887322 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887332 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887342 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887353 4862 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887362 4862 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887374 4862 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887384 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887394 4862 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887404 4862 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887413 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887423 4862 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887433 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887446 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887459 4862 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887473 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887486 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887500 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887511 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887521 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887531 4862 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887540 4862 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.887551 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.923037 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885456 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885462 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.885484 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.886031 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.894259 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.894435 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.894614 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.894980 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.895031 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.895241 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.895811 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.895840 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.895870 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.896171 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.896230 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.896246 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.903017 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.908220 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.908633 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.910602 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.912731 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.913185 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.913623 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.913973 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.914064 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.914321 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.914484 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.914669 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.916523 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.917035 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.917315 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.918254 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.919356 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.919719 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.920819 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.921478 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.921881 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.921969 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.922368 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.922812 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.922836 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.922848 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.923763 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.923902 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.923971 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.924260 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.924447 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.924566 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.924602 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.925467 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.925590 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.925894 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.925934 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.926203 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.926286 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.926608 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.926832 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.926889 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.926907 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.927040 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.929426 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.930227 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.931513 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.938545 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.938637 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.938656 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.938803 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.938839 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.938933 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.938962 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.939152 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.939319 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.939397 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.939539 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.939562 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.939722 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.939891 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.939944 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.940125 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.940181 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.941338 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.940110 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.940385 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.940371 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.940426 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.940697 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.940754 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.940988 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.941191 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.941530 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.941604 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.943015 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.943027 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.943313 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.943661 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.944281 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.944515 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.944741 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.944772 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.945273 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.945416 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.945539 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.945901 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.946172 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.946337 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.946453 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.946569 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.947060 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.947393 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.947619 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.948598 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.949051 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.949449 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.949693 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.950000 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.950629 4862 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.950719 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.950743 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:22.450719304 +0000 UTC m=+19.983233225 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.950929 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.951114 4862 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.951624 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:22.451573969 +0000 UTC m=+19.984087890 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.951340 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.952005 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.952351 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.952464 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.952918 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.952980 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.953204 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.953265 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.953408 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.953557 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.953601 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.953623 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.953629 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.953645 4862 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.953938 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.954053 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.954084 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:22.454065651 +0000 UTC m=+19.986579572 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.953080 4862 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.953995 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.955443 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.958764 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.959411 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.959548 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.964998 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.965397 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.965490 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.965714 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.969829 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.973418 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.974055 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.976187 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.976491 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82"} Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.976500 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.976514 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.976508 4862 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82" exitCode=255 Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.979043 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.979214 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.982572 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.991545 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.993394 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.993425 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.993443 4862 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:21 crc kubenswrapper[4862]: E1128 11:22:21.993514 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:22.493490344 +0000 UTC m=+20.026004265 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.994403 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.994823 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.996250 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997312 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997371 4862 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997385 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997421 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997435 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997451 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997461 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997473 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997505 4862 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997520 4862 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997532 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997542 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997555 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997584 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997596 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997607 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997620 4862 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997632 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997660 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997669 4862 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997682 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997691 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997700 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997710 4862 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997741 4862 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997750 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997761 4862 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997776 4862 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997787 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997815 4862 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997825 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997840 4862 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997851 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997862 4862 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997910 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997924 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997935 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997945 4862 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997976 4862 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997986 4862 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997996 4862 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.998006 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.998018 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.998065 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.998077 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:21 crc kubenswrapper[4862]: I1128 11:22:21.997346 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.000530 4862 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.009577 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.009730 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.009794 4862 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.009863 4862 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.009919 4862 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.009973 4862 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010034 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010123 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010181 4862 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010246 4862 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010306 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010370 4862 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010430 4862 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010485 4862 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010542 4862 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010601 4862 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010659 4862 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010718 4862 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010771 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010843 4862 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010951 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011008 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011060 4862 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011180 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011249 4862 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011301 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011365 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011431 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011490 4862 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011545 4862 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011607 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011667 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011722 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.011981 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012055 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012154 4862 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012219 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012289 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012346 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012637 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012708 4862 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012771 4862 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012829 4862 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012899 4862 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.012953 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013005 4862 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013067 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013162 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013232 4862 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013291 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013349 4862 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013407 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013459 4862 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013513 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013577 4862 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013639 4862 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013696 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013753 4862 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013810 4862 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013868 4862 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013929 4862 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.013982 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014036 4862 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014107 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014173 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014343 4862 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014408 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014462 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014523 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014580 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014644 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014707 4862 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014760 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014814 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014873 4862 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014929 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.014994 4862 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015055 4862 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015128 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015195 4862 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015252 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015313 4862 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015392 4862 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015449 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015503 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015564 4862 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015623 4862 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015681 4862 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015751 4862 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015812 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015875 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015931 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015989 4862 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.016047 4862 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.016165 4862 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.016228 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:21.996495 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.012611 4862 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.009208 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.009511 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.015293 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.010882 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.011935 4862 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.017636 4862 scope.go:117] "RemoveContainer" containerID="a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.024406 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.026497 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.039922 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.073919 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.076821 4862 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.081839 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.089786 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.096702 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.097557 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: W1128 11:22:22.117133 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-c9482f6a2a6708fd03de85ec81a367d8dd7f73cf27f30dd06f6d82b2baa659e9 WatchSource:0}: Error finding container c9482f6a2a6708fd03de85ec81a367d8dd7f73cf27f30dd06f6d82b2baa659e9: Status 404 returned error can't find the container with id c9482f6a2a6708fd03de85ec81a367d8dd7f73cf27f30dd06f6d82b2baa659e9 Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.117514 4862 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.117541 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.117554 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.117567 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.117578 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.122309 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.133795 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.145234 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.157547 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.169922 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.191195 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.194678 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.216537 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.230176 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.239444 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.421313 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.421526 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:22:23.421492107 +0000 UTC m=+20.954006038 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.522764 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.522807 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.522827 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.522847 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.522946 4862 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.522963 4862 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523000 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:23.522986939 +0000 UTC m=+21.055500860 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523043 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:23.52302581 +0000 UTC m=+21.055539731 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523128 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523177 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523177 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523192 4862 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523196 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523211 4862 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523242 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:23.523235326 +0000 UTC m=+21.055749247 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:22 crc kubenswrapper[4862]: E1128 11:22:22.523288 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:23.523264417 +0000 UTC m=+21.055778338 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.684904 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-kqpd4"] Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.685284 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-25clg"] Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.685418 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.685449 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-2nh6f"] Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.685528 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-25clg" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.686631 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-fm8j2"] Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.686953 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.692637 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.692996 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.693358 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.695022 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.695404 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.700779 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.700921 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.701060 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.700784 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.701373 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.701384 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.701425 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.701447 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.701536 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.701641 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.701646 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.722855 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.736164 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.781434 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.802277 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.822139 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827735 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fc6de873-3ea5-4223-a260-9451e5e6e803-hosts-file\") pod \"node-resolver-25clg\" (UID: \"fc6de873-3ea5-4223-a260-9451e5e6e803\") " pod="openshift-dns/node-resolver-25clg" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827771 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-daemon-config\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827800 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-run-netns\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827822 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827855 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a6845439-1a4c-4f28-9ba4-29149d545995-proxy-tls\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827873 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-cni-dir\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827893 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-var-lib-cni-bin\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827917 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-cnibin\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827950 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6845439-1a4c-4f28-9ba4-29149d545995-mcd-auth-proxy-config\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827973 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-etc-kubernetes\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.827995 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-system-cni-dir\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828038 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828062 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-os-release\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828112 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-run-k8s-cni-cncf-io\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828137 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59vmn\" (UniqueName: \"kubernetes.io/projected/d6abffd5-6086-4a32-ba8d-e7c669548c5f-kube-api-access-59vmn\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828171 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2p4r\" (UniqueName: \"kubernetes.io/projected/a6845439-1a4c-4f28-9ba4-29149d545995-kube-api-access-f2p4r\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828196 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-hostroot\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828227 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-cni-binary-copy\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828249 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a6845439-1a4c-4f28-9ba4-29149d545995-rootfs\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828270 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-socket-dir-parent\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828293 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-var-lib-kubelet\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828313 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-run-multus-certs\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828334 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zlwp\" (UniqueName: \"kubernetes.io/projected/fc6de873-3ea5-4223-a260-9451e5e6e803-kube-api-access-8zlwp\") pod \"node-resolver-25clg\" (UID: \"fc6de873-3ea5-4223-a260-9451e5e6e803\") " pod="openshift-dns/node-resolver-25clg" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828359 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-cnibin\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828380 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d6abffd5-6086-4a32-ba8d-e7c669548c5f-cni-binary-copy\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828402 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-var-lib-cni-multus\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828425 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-system-cni-dir\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828449 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-os-release\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828469 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq7pb\" (UniqueName: \"kubernetes.io/projected/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-kube-api-access-sq7pb\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.828490 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-conf-dir\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.839055 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.841944 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.842460 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.843311 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.843886 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.844473 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.844990 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.845612 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.846184 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.846767 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.847278 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.847752 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.848413 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.848878 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.849408 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.849897 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.850417 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.850934 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.853969 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.854526 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.855105 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.855923 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.856460 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.857253 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.857857 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.858260 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.858397 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.859324 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.860294 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.860750 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.861401 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.862358 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.862810 4862 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.862909 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.864871 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.865415 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.865840 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.867717 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.868331 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.868798 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.869740 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.870376 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.871201 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.871765 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.872704 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.873635 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.874071 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.874999 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.875492 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.876553 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.876473 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.877000 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.877468 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.878283 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.878779 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.879722 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.880186 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.893688 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.906918 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.921110 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929673 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-run-netns\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929715 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929747 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a6845439-1a4c-4f28-9ba4-29149d545995-proxy-tls\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929789 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-cni-dir\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929797 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-run-netns\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929810 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-var-lib-cni-bin\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929826 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-etc-kubernetes\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929843 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-cnibin\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929889 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6845439-1a4c-4f28-9ba4-29149d545995-mcd-auth-proxy-config\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929906 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-os-release\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929941 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-run-k8s-cni-cncf-io\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929960 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59vmn\" (UniqueName: \"kubernetes.io/projected/d6abffd5-6086-4a32-ba8d-e7c669548c5f-kube-api-access-59vmn\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929972 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-var-lib-cni-bin\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.929978 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-system-cni-dir\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930025 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-system-cni-dir\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930124 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-run-k8s-cni-cncf-io\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930127 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-cni-dir\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930141 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-etc-kubernetes\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930193 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-cnibin\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930243 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930303 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-hostroot\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930355 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-hostroot\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930395 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-os-release\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930471 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2p4r\" (UniqueName: \"kubernetes.io/projected/a6845439-1a4c-4f28-9ba4-29149d545995-kube-api-access-f2p4r\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930515 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-cni-binary-copy\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930538 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a6845439-1a4c-4f28-9ba4-29149d545995-rootfs\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930562 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-socket-dir-parent\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930766 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930788 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-var-lib-kubelet\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930773 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a6845439-1a4c-4f28-9ba4-29149d545995-rootfs\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930829 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-socket-dir-parent\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930890 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zlwp\" (UniqueName: \"kubernetes.io/projected/fc6de873-3ea5-4223-a260-9451e5e6e803-kube-api-access-8zlwp\") pod \"node-resolver-25clg\" (UID: \"fc6de873-3ea5-4223-a260-9451e5e6e803\") " pod="openshift-dns/node-resolver-25clg" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930909 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-var-lib-kubelet\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930958 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-cnibin\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.930978 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6845439-1a4c-4f28-9ba4-29149d545995-mcd-auth-proxy-config\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931066 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d6abffd5-6086-4a32-ba8d-e7c669548c5f-cni-binary-copy\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931081 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-cnibin\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931126 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-var-lib-cni-multus\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931158 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-var-lib-cni-multus\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931194 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-run-multus-certs\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931227 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-host-run-multus-certs\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931238 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-cni-binary-copy\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931262 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-system-cni-dir\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931294 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-os-release\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931322 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq7pb\" (UniqueName: \"kubernetes.io/projected/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-kube-api-access-sq7pb\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931348 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-conf-dir\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931357 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-system-cni-dir\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931373 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-daemon-config\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931365 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-os-release\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931400 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fc6de873-3ea5-4223-a260-9451e5e6e803-hosts-file\") pod \"node-resolver-25clg\" (UID: \"fc6de873-3ea5-4223-a260-9451e5e6e803\") " pod="openshift-dns/node-resolver-25clg" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931409 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-conf-dir\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931337 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931479 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fc6de873-3ea5-4223-a260-9451e5e6e803-hosts-file\") pod \"node-resolver-25clg\" (UID: \"fc6de873-3ea5-4223-a260-9451e5e6e803\") " pod="openshift-dns/node-resolver-25clg" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.931654 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d6abffd5-6086-4a32-ba8d-e7c669548c5f-cni-binary-copy\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.932048 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d6abffd5-6086-4a32-ba8d-e7c669548c5f-multus-daemon-config\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.935755 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a6845439-1a4c-4f28-9ba4-29149d545995-proxy-tls\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.957673 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2p4r\" (UniqueName: \"kubernetes.io/projected/a6845439-1a4c-4f28-9ba4-29149d545995-kube-api-access-f2p4r\") pod \"machine-config-daemon-fm8j2\" (UID: \"a6845439-1a4c-4f28-9ba4-29149d545995\") " pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.961371 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zlwp\" (UniqueName: \"kubernetes.io/projected/fc6de873-3ea5-4223-a260-9451e5e6e803-kube-api-access-8zlwp\") pod \"node-resolver-25clg\" (UID: \"fc6de873-3ea5-4223-a260-9451e5e6e803\") " pod="openshift-dns/node-resolver-25clg" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.961955 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq7pb\" (UniqueName: \"kubernetes.io/projected/68a1bd24-7d7d-4d41-9f82-ff6158c145d0-kube-api-access-sq7pb\") pod \"multus-additional-cni-plugins-2nh6f\" (UID: \"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\") " pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.966229 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59vmn\" (UniqueName: \"kubernetes.io/projected/d6abffd5-6086-4a32-ba8d-e7c669548c5f-kube-api-access-59vmn\") pod \"multus-kqpd4\" (UID: \"d6abffd5-6086-4a32-ba8d-e7c669548c5f\") " pod="openshift-multus/multus-kqpd4" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.966176 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.983238 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.985306 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b"} Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.985607 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.989845 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a9eb5919875a618cf2e64745e22d1a652ec65aadd1c4a260da9593da128bd9b3"} Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.991031 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e"} Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.991055 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c9482f6a2a6708fd03de85ec81a367d8dd7f73cf27f30dd06f6d82b2baa659e9"} Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.991611 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.993338 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e"} Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.993365 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660"} Nov 28 11:22:22 crc kubenswrapper[4862]: I1128 11:22:22.993375 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"48f9a289b96d420bd471efb14130014a7bd4fafbd248a95c6a33545574ec6a83"} Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.003168 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-kqpd4" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.007702 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.012006 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" Nov 28 11:22:23 crc kubenswrapper[4862]: W1128 11:22:23.013758 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6abffd5_6086_4a32_ba8d_e7c669548c5f.slice/crio-ceac2fb9838eee7c24bba573e7a87f54f8ce64b852b59b587a6822675370a2b1 WatchSource:0}: Error finding container ceac2fb9838eee7c24bba573e7a87f54f8ce64b852b59b587a6822675370a2b1: Status 404 returned error can't find the container with id ceac2fb9838eee7c24bba573e7a87f54f8ce64b852b59b587a6822675370a2b1 Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.018908 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-25clg" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.028922 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.036472 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.058860 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.082076 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.091078 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ccpp6"] Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.091809 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.098054 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.098136 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.098266 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.098268 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.099885 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.100078 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.100281 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.100349 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.124781 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.140898 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.168454 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.195122 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.210671 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.229623 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236473 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-slash\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236510 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9916d783-c2df-47fa-9c38-ee7537db24d4-ovn-node-metrics-cert\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236526 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-netns\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236542 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-etc-openvswitch\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236559 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-netd\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236585 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-log-socket\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236607 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-ovn\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236622 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-kubelet\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236638 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-systemd\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236659 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236675 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-var-lib-openvswitch\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236703 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-config\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236722 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjwcr\" (UniqueName: \"kubernetes.io/projected/9916d783-c2df-47fa-9c38-ee7537db24d4-kube-api-access-tjwcr\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236743 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-openvswitch\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236771 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-bin\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236796 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-systemd-units\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236823 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-env-overrides\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236855 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-node-log\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236870 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-ovn-kubernetes\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.236903 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-script-lib\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.259554 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.273240 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.289166 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.307040 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.324051 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337557 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-log-socket\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337618 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-ovn\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337650 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-kubelet\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337674 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-systemd\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337695 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337719 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-var-lib-openvswitch\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337739 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-config\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337758 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjwcr\" (UniqueName: \"kubernetes.io/projected/9916d783-c2df-47fa-9c38-ee7537db24d4-kube-api-access-tjwcr\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337774 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-openvswitch\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337793 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-bin\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337818 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-systemd-units\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337835 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-env-overrides\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337863 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-node-log\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337879 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-ovn-kubernetes\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337896 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-script-lib\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337915 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-slash\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337933 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9916d783-c2df-47fa-9c38-ee7537db24d4-ovn-node-metrics-cert\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337953 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-netns\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337971 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-etc-openvswitch\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.337988 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-netd\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338067 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-netd\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338134 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-log-socket\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338159 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-ovn\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338181 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-kubelet\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338204 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-systemd\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338228 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338258 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-var-lib-openvswitch\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338404 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-node-log\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338716 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-bin\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338789 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-systemd-units\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338874 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-config\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338886 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-ovn-kubernetes\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338927 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-etc-openvswitch\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.338886 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-netns\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.339362 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-openvswitch\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.339468 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-slash\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.339493 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-script-lib\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.342486 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-env-overrides\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.349743 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9916d783-c2df-47fa-9c38-ee7537db24d4-ovn-node-metrics-cert\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.372596 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjwcr\" (UniqueName: \"kubernetes.io/projected/9916d783-c2df-47fa-9c38-ee7537db24d4-kube-api-access-tjwcr\") pod \"ovnkube-node-ccpp6\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.382994 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.413793 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.416295 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.438714 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.438955 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:22:25.438938884 +0000 UTC m=+22.971452805 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.447689 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.471733 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.498742 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.514811 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.529917 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.539247 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.539404 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.539502 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.539603 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.539452 4862 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.539792 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:25.539775826 +0000 UTC m=+23.072289747 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.539533 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.539951 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.540013 4862 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.540113 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:25.540104725 +0000 UTC m=+23.072618636 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.539590 4862 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.540244 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:25.540236409 +0000 UTC m=+23.072750320 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.539670 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.540369 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.540427 4862 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.540507 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:25.540499327 +0000 UTC m=+23.073013248 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.547494 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.553856 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.563451 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.567490 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.575042 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.577909 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.594902 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.612223 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.633953 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.654610 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.678228 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.700297 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.720081 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.739774 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.757990 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.778351 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.801266 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.820074 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.834700 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.837949 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.838222 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.838415 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.838620 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.838617 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:23 crc kubenswrapper[4862]: E1128 11:22:23.838720 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.854751 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.872606 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.892400 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.921934 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.966211 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.997125 4862 generic.go:334] "Generic (PLEG): container finished" podID="68a1bd24-7d7d-4d41-9f82-ff6158c145d0" containerID="fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276" exitCode=0 Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.997203 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" event={"ID":"68a1bd24-7d7d-4d41-9f82-ff6158c145d0","Type":"ContainerDied","Data":"fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276"} Nov 28 11:22:23 crc kubenswrapper[4862]: I1128 11:22:23.997249 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" event={"ID":"68a1bd24-7d7d-4d41-9f82-ff6158c145d0","Type":"ContainerStarted","Data":"4b6efa7b58b84359655d787c322aaffae8b921f9839791f32ecdefd9c045a93c"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.007144 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69" exitCode=0 Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.007245 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.007278 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"7a759715337151e5488c580bd6e803f7ea34cd6c692e8c9775e9ab62a62bdce5"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.009772 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.010309 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.010323 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"05da2155d9fc36c5184e3cc03af30eb8ff74153c85a85e819864c29bfa9caed0"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.011033 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kqpd4" event={"ID":"d6abffd5-6086-4a32-ba8d-e7c669548c5f","Type":"ContainerStarted","Data":"3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.011122 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kqpd4" event={"ID":"d6abffd5-6086-4a32-ba8d-e7c669548c5f","Type":"ContainerStarted","Data":"ceac2fb9838eee7c24bba573e7a87f54f8ce64b852b59b587a6822675370a2b1"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.023204 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-25clg" event={"ID":"fc6de873-3ea5-4223-a260-9451e5e6e803","Type":"ContainerStarted","Data":"cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.023244 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-25clg" event={"ID":"fc6de873-3ea5-4223-a260-9451e5e6e803","Type":"ContainerStarted","Data":"942fe024f5aa93cc2fea135f3aadaf643f52176b4f8d97e96c50ab8398867a7a"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.025161 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.095959 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.200519 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.250835 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.266810 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.283984 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.302389 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.324435 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.348778 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.367445 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.380528 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.397604 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.416718 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.431863 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.443623 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.459586 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.474911 4862 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.477219 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.477664 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.477675 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.477793 4862 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.479646 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.485411 4862 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.485729 4862 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.487012 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.487043 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.487056 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.487070 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.487081 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:24Z","lastTransitionTime":"2025-11-28T11:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.496277 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: E1128 11:22:24.507568 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.511231 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.511264 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.511274 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.511290 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.511299 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:24Z","lastTransitionTime":"2025-11-28T11:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:24 crc kubenswrapper[4862]: E1128 11:22:24.527012 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.531537 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.531560 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.531569 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.531586 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.531597 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:24Z","lastTransitionTime":"2025-11-28T11:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:24 crc kubenswrapper[4862]: E1128 11:22:24.544848 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.553561 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.553606 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.553619 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.553638 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.553650 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:24Z","lastTransitionTime":"2025-11-28T11:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:24 crc kubenswrapper[4862]: E1128 11:22:24.566970 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.570959 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.571039 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.571104 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.571122 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.571133 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:24Z","lastTransitionTime":"2025-11-28T11:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:24 crc kubenswrapper[4862]: E1128 11:22:24.585620 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: E1128 11:22:24.585747 4862 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.592168 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.592213 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.592229 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.592250 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.592266 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:24Z","lastTransitionTime":"2025-11-28T11:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.645822 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-tpk22"] Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.646406 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.649713 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.650443 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.650651 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.653240 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.662611 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.684212 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.702491 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.705780 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.705822 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.705836 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.705860 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.705874 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:24Z","lastTransitionTime":"2025-11-28T11:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.719447 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.736176 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.749913 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.764341 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b99r6\" (UniqueName: \"kubernetes.io/projected/4f84710d-eb96-4879-8704-1e8dda17d4b0-kube-api-access-b99r6\") pod \"node-ca-tpk22\" (UID: \"4f84710d-eb96-4879-8704-1e8dda17d4b0\") " pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.764391 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4f84710d-eb96-4879-8704-1e8dda17d4b0-serviceca\") pod \"node-ca-tpk22\" (UID: \"4f84710d-eb96-4879-8704-1e8dda17d4b0\") " pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.764444 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f84710d-eb96-4879-8704-1e8dda17d4b0-host\") pod \"node-ca-tpk22\" (UID: \"4f84710d-eb96-4879-8704-1e8dda17d4b0\") " pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.776267 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.808335 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.808397 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.808407 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.808432 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.808444 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:24Z","lastTransitionTime":"2025-11-28T11:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.820971 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.858950 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.865600 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f84710d-eb96-4879-8704-1e8dda17d4b0-host\") pod \"node-ca-tpk22\" (UID: \"4f84710d-eb96-4879-8704-1e8dda17d4b0\") " pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.865672 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b99r6\" (UniqueName: \"kubernetes.io/projected/4f84710d-eb96-4879-8704-1e8dda17d4b0-kube-api-access-b99r6\") pod \"node-ca-tpk22\" (UID: \"4f84710d-eb96-4879-8704-1e8dda17d4b0\") " pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.865702 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4f84710d-eb96-4879-8704-1e8dda17d4b0-serviceca\") pod \"node-ca-tpk22\" (UID: \"4f84710d-eb96-4879-8704-1e8dda17d4b0\") " pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.865757 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f84710d-eb96-4879-8704-1e8dda17d4b0-host\") pod \"node-ca-tpk22\" (UID: \"4f84710d-eb96-4879-8704-1e8dda17d4b0\") " pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.866958 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4f84710d-eb96-4879-8704-1e8dda17d4b0-serviceca\") pod \"node-ca-tpk22\" (UID: \"4f84710d-eb96-4879-8704-1e8dda17d4b0\") " pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.910155 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b99r6\" (UniqueName: \"kubernetes.io/projected/4f84710d-eb96-4879-8704-1e8dda17d4b0-kube-api-access-b99r6\") pod \"node-ca-tpk22\" (UID: \"4f84710d-eb96-4879-8704-1e8dda17d4b0\") " pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.912185 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.912245 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.912258 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.912278 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.912291 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:24Z","lastTransitionTime":"2025-11-28T11:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.918462 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.961303 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tpk22" Nov 28 11:22:24 crc kubenswrapper[4862]: I1128 11:22:24.963976 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:24Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:24 crc kubenswrapper[4862]: W1128 11:22:24.978666 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f84710d_eb96_4879_8704_1e8dda17d4b0.slice/crio-50f3c0040c98085a470dd22a36016c97cb35881b729ca6dcba8ca486ea3471da WatchSource:0}: Error finding container 50f3c0040c98085a470dd22a36016c97cb35881b729ca6dcba8ca486ea3471da: Status 404 returned error can't find the container with id 50f3c0040c98085a470dd22a36016c97cb35881b729ca6dcba8ca486ea3471da Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.002249 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.015895 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.015935 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.015945 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.015977 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.015995 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.029960 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.032973 4862 generic.go:334] "Generic (PLEG): container finished" podID="68a1bd24-7d7d-4d41-9f82-ff6158c145d0" containerID="6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c" exitCode=0 Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.033049 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" event={"ID":"68a1bd24-7d7d-4d41-9f82-ff6158c145d0","Type":"ContainerDied","Data":"6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.034945 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tpk22" event={"ID":"4f84710d-eb96-4879-8704-1e8dda17d4b0","Type":"ContainerStarted","Data":"50f3c0040c98085a470dd22a36016c97cb35881b729ca6dcba8ca486ea3471da"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.039878 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.039972 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.039996 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.040019 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.040039 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.051356 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.078227 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.114428 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.119111 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.119155 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.119168 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.119188 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.119201 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.155742 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.193882 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.224282 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.224349 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.224366 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.224390 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.224407 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.234662 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.281141 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.321720 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.327449 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.327589 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.327662 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.327733 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.327803 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.355140 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.399343 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.430542 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.430801 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.430875 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.430944 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.431010 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.440016 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.472373 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.472578 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:22:29.472549138 +0000 UTC m=+27.005063059 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.477944 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.530348 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.541711 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.541770 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.541783 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.541801 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.541815 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.554139 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.573380 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.573581 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.573598 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.573587 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.573609 4862 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.573768 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.573830 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:29.573807552 +0000 UTC m=+27.106321473 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.573873 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.573990 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.574024 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.574047 4862 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.574002 4862 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.574133 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:29.574123761 +0000 UTC m=+27.106637682 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.574150 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:29.574141572 +0000 UTC m=+27.106655493 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.573638 4862 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.574225 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:29.574193863 +0000 UTC m=+27.106707824 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.620664 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.639391 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.644415 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.644453 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.644463 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.644478 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.644493 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.680324 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.716445 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:25Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.748068 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.748127 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.748137 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.748165 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.748177 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.837846 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.837901 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.837950 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.838078 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.838249 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:25 crc kubenswrapper[4862]: E1128 11:22:25.838369 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.851251 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.851439 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.851539 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.851619 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.851694 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.954271 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.954740 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.954752 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.954770 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:25 crc kubenswrapper[4862]: I1128 11:22:25.954782 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:25Z","lastTransitionTime":"2025-11-28T11:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.046725 4862 generic.go:334] "Generic (PLEG): container finished" podID="68a1bd24-7d7d-4d41-9f82-ff6158c145d0" containerID="7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09" exitCode=0 Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.046783 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" event={"ID":"68a1bd24-7d7d-4d41-9f82-ff6158c145d0","Type":"ContainerDied","Data":"7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.053522 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.057478 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.057552 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.057563 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.057581 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.057594 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:26Z","lastTransitionTime":"2025-11-28T11:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.058560 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tpk22" event={"ID":"4f84710d-eb96-4879-8704-1e8dda17d4b0","Type":"ContainerStarted","Data":"d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.075960 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.088871 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.112602 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.131828 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.148762 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.164263 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.164301 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.164310 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.164325 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.164358 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:26Z","lastTransitionTime":"2025-11-28T11:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.176547 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.199679 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.216926 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.230578 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.253345 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.268900 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.270202 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.270265 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.270280 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.270306 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.270324 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:26Z","lastTransitionTime":"2025-11-28T11:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.298405 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.317348 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.332815 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.353377 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.369641 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.373921 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.373977 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.373987 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.374010 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.374022 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:26Z","lastTransitionTime":"2025-11-28T11:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.395896 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.447212 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.477800 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.477867 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.477884 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.477907 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.477923 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:26Z","lastTransitionTime":"2025-11-28T11:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.479990 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.520771 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.557311 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.581513 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.581891 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.581999 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.582074 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.582177 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:26Z","lastTransitionTime":"2025-11-28T11:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.597886 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.636806 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.679434 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.685785 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.685854 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.685874 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.685901 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.685920 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:26Z","lastTransitionTime":"2025-11-28T11:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.721938 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.756776 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.789209 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.789273 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.789288 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.789310 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.789326 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:26Z","lastTransitionTime":"2025-11-28T11:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.803041 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.835305 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.875328 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.897933 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.897983 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.898036 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.898057 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.898077 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:26Z","lastTransitionTime":"2025-11-28T11:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:26 crc kubenswrapper[4862]: I1128 11:22:26.926054 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.000366 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.000416 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.000426 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.000444 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.000455 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.065160 4862 generic.go:334] "Generic (PLEG): container finished" podID="68a1bd24-7d7d-4d41-9f82-ff6158c145d0" containerID="f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6" exitCode=0 Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.065527 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" event={"ID":"68a1bd24-7d7d-4d41-9f82-ff6158c145d0","Type":"ContainerDied","Data":"f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.083660 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.105402 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.106486 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.106546 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.106560 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.106581 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.106594 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.121312 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.135706 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.153286 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.164234 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.197575 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.210625 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.210673 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.210687 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.210707 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.210720 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.234710 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.277532 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.315076 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.315141 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.315153 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.315172 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.315183 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.321500 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.364809 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.399813 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.418907 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.418956 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.418965 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.418982 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.418995 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.439990 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.482817 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.519510 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:27Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.521471 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.521508 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.521518 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.521537 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.521552 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.624346 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.624403 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.624416 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.624436 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.624446 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.727293 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.727389 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.727408 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.727436 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.727455 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.831975 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.832068 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.832121 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.832157 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.832182 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.838647 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.838683 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.838677 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:27 crc kubenswrapper[4862]: E1128 11:22:27.839085 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:27 crc kubenswrapper[4862]: E1128 11:22:27.838863 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:27 crc kubenswrapper[4862]: E1128 11:22:27.839279 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.935207 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.935264 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.935278 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.935305 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:27 crc kubenswrapper[4862]: I1128 11:22:27.935321 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:27Z","lastTransitionTime":"2025-11-28T11:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.038478 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.038533 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.038546 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.038566 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.038579 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.073682 4862 generic.go:334] "Generic (PLEG): container finished" podID="68a1bd24-7d7d-4d41-9f82-ff6158c145d0" containerID="ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd" exitCode=0 Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.073784 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" event={"ID":"68a1bd24-7d7d-4d41-9f82-ff6158c145d0","Type":"ContainerDied","Data":"ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.078459 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.091301 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.115271 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.141142 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.142079 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.142151 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.142166 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.142185 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.142199 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.169701 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.193842 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.211166 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.229966 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.246474 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.246560 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.246578 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.246603 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.246624 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.249178 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.261704 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.278602 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.296543 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.312404 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.329612 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.342709 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.349360 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.349430 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.349450 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.349475 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.349493 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.360157 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:28Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.452617 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.452772 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.452786 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.452805 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.452822 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.556087 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.556337 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.556403 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.556439 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.556460 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.659619 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.659710 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.659737 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.659769 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.659797 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.763258 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.763330 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.763349 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.763378 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.763395 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.866674 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.866752 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.866775 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.866804 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.866825 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.970388 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.970462 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.970479 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.970506 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:28 crc kubenswrapper[4862]: I1128 11:22:28.970524 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:28Z","lastTransitionTime":"2025-11-28T11:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.072940 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.073010 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.073031 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.073055 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.073073 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:29Z","lastTransitionTime":"2025-11-28T11:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.177256 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.177338 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.177355 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.177405 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.177424 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:29Z","lastTransitionTime":"2025-11-28T11:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.280772 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.280830 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.280850 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.280878 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.280897 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:29Z","lastTransitionTime":"2025-11-28T11:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.384620 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.384685 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.384706 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.384743 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.384769 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:29Z","lastTransitionTime":"2025-11-28T11:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.489744 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.490318 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.490343 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.490381 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.490406 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:29Z","lastTransitionTime":"2025-11-28T11:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.521752 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.522061 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:22:37.522026962 +0000 UTC m=+35.054540923 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.594146 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.594196 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.594214 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.594240 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.594262 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:29Z","lastTransitionTime":"2025-11-28T11:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.623161 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.623239 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.623309 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.623368 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623485 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623506 4862 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623540 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623588 4862 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623665 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623801 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623673 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:37.623644917 +0000 UTC m=+35.156158878 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623506 4862 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623935 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:37.623895605 +0000 UTC m=+35.156409576 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.623990 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:37.623956916 +0000 UTC m=+35.156470877 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.624031 4862 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.624190 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:37.624173763 +0000 UTC m=+35.156687724 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.699398 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.699508 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.699527 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.699598 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.699621 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:29Z","lastTransitionTime":"2025-11-28T11:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.802789 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.802855 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.802873 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.802896 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.802916 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:29Z","lastTransitionTime":"2025-11-28T11:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.838153 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.838257 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.838305 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.838444 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.838578 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:29 crc kubenswrapper[4862]: E1128 11:22:29.838689 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.905302 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.905721 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.905796 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.905911 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:29 crc kubenswrapper[4862]: I1128 11:22:29.905983 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:29Z","lastTransitionTime":"2025-11-28T11:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.008766 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.008830 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.008849 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.008874 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.008893 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.098353 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.098804 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.103854 4862 generic.go:334] "Generic (PLEG): container finished" podID="68a1bd24-7d7d-4d41-9f82-ff6158c145d0" containerID="33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c" exitCode=0 Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.103916 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" event={"ID":"68a1bd24-7d7d-4d41-9f82-ff6158c145d0","Type":"ContainerDied","Data":"33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.111617 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.111850 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.112003 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.112200 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.112336 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.117664 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.136975 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.146738 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.152664 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.196806 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.216069 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.216122 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.216134 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.216155 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.216169 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.216616 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.236252 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.256296 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.271954 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.285005 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.296612 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.311187 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.319676 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.319711 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.319725 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.319744 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.319758 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.332571 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.354125 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.371812 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.385738 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.400629 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.421964 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.422005 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.422014 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.422034 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.422046 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.423894 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.438198 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.452158 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.462203 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.475370 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.490254 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.510366 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.528594 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.528645 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.528660 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.528681 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.528700 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.544671 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.559303 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.576048 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.591294 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.605341 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.617852 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.631824 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.631867 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.631880 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.631904 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.631919 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.637060 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:30Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.735444 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.735550 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.735571 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.735636 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.735659 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.839391 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.839472 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.839496 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.839527 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.839549 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.943071 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.943146 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.943156 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.943175 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:30 crc kubenswrapper[4862]: I1128 11:22:30.943188 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:30Z","lastTransitionTime":"2025-11-28T11:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.046546 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.046626 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.046649 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.047065 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.047176 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.116569 4862 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.117060 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" event={"ID":"68a1bd24-7d7d-4d41-9f82-ff6158c145d0","Type":"ContainerStarted","Data":"62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.117438 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.138649 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.149314 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.150680 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.150723 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.150739 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.150760 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.150781 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.163236 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.185145 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.209812 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.235771 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.255849 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.255885 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.255898 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.255916 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.255930 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.257749 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.271667 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.284656 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.298520 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.324891 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.346789 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.359018 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.359085 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.359129 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.359158 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.359176 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.365973 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.378845 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.394163 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.404663 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.423501 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.439251 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.452912 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.462161 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.462340 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.462401 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.462492 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.462616 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.465481 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.487497 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.503224 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.517044 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.530448 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.545704 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.558586 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.564953 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.565164 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.565295 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.565413 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.565521 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.574836 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.589615 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.607719 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.631853 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.647934 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:31Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.667964 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.668004 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.668015 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.668033 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.668047 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.771133 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.771195 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.771214 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.771239 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.771256 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.837921 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:31 crc kubenswrapper[4862]: E1128 11:22:31.838236 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.838828 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:31 crc kubenswrapper[4862]: E1128 11:22:31.838933 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.839020 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:31 crc kubenswrapper[4862]: E1128 11:22:31.839140 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.874113 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.874152 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.874161 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.874178 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.874188 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.977396 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.977462 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.977474 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.977493 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:31 crc kubenswrapper[4862]: I1128 11:22:31.977505 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:31Z","lastTransitionTime":"2025-11-28T11:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.081572 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.081618 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.081628 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.081649 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.081673 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:32Z","lastTransitionTime":"2025-11-28T11:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.120358 4862 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.184557 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.184613 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.184626 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.184654 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.184668 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:32Z","lastTransitionTime":"2025-11-28T11:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.287265 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.287311 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.287324 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.287343 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.287358 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:32Z","lastTransitionTime":"2025-11-28T11:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.390343 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.390392 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.390405 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.390426 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.390450 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:32Z","lastTransitionTime":"2025-11-28T11:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.492729 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.492800 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.492820 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.492846 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.492868 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:32Z","lastTransitionTime":"2025-11-28T11:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.596583 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.596653 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.596665 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.596700 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.596715 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:32Z","lastTransitionTime":"2025-11-28T11:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.702164 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.702663 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.702730 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.702827 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.702887 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:32Z","lastTransitionTime":"2025-11-28T11:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.805971 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.806047 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.806148 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.806187 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.806209 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:32Z","lastTransitionTime":"2025-11-28T11:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.858444 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:32Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.880254 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:32Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.894923 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:32Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.909302 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.909350 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.909369 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.909394 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.909413 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:32Z","lastTransitionTime":"2025-11-28T11:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.914497 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:32Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.931402 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:32Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.948361 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:32Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:32 crc kubenswrapper[4862]: I1128 11:22:32.981566 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:32Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.008365 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:33Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.011628 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.011678 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.011698 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.011722 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.011739 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.027976 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:33Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.049853 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:33Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.065407 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:33Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.077387 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:33Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.114456 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.114501 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.114512 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.114528 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.114540 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.123400 4862 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.144236 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:33Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.162690 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:33Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.181824 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:33Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.216740 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.216772 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.216781 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.216794 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.216804 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.320310 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.320373 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.320388 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.320410 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.320426 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.422758 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.422823 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.422844 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.422902 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.422920 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.525871 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.525929 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.525944 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.525967 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.525988 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.628829 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.628926 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.628940 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.628962 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.628976 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.732652 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.732721 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.732740 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.732766 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.732785 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.836573 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.836631 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.836650 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.836677 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.836697 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.837906 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:33 crc kubenswrapper[4862]: E1128 11:22:33.838053 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.837905 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.838140 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:33 crc kubenswrapper[4862]: E1128 11:22:33.838235 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:33 crc kubenswrapper[4862]: E1128 11:22:33.838566 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.939738 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.939793 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.939818 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.939848 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:33 crc kubenswrapper[4862]: I1128 11:22:33.939870 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:33Z","lastTransitionTime":"2025-11-28T11:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.044124 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.044209 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.044234 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.044268 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.044292 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.132862 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/0.log" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.137845 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9" exitCode=1 Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.137930 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.139065 4862 scope.go:117] "RemoveContainer" containerID="17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.146581 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.146646 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.146665 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.146693 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.146713 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.161746 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.185482 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.210917 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.227478 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.245746 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.250296 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.250360 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.250376 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.250402 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.250424 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.263503 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.289078 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.318076 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:33Z\\\",\\\"message\\\":\\\"7973 6144 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1128 11:22:32.838039 6144 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 11:22:32.837333 6144 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:32.838567 6144 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:32.838589 6144 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:32.838611 6144 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 11:22:32.838634 6144 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 11:22:32.838649 6144 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 11:22:32.838647 6144 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 11:22:32.838672 6144 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 11:22:32.838679 6144 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 11:22:32.838724 6144 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 11:22:32.838735 6144 factory.go:656] Stopping watch factory\\\\nI1128 11:22:32.838802 6144 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.339631 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.354322 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.354399 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.354420 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.354447 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.354467 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.356086 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.371693 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.388980 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.402768 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.429001 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.449889 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.458145 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.458226 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.458241 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.458262 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.458278 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.561692 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.561754 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.561763 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.561784 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.561799 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.665331 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.665757 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.665982 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.666189 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.666355 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.770032 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.770453 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.770473 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.770496 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.770518 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.772041 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.772131 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.772154 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.772179 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.772198 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: E1128 11:22:34.792924 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.798822 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.799162 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.799403 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.799581 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.799733 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: E1128 11:22:34.821862 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.828718 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.828771 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.828827 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.828859 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.828883 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: E1128 11:22:34.853212 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.858323 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.858382 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.858403 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.858426 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.858443 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: E1128 11:22:34.878175 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.884562 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.884619 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.884639 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.884661 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.884682 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.898944 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn"] Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.899628 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.903190 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 11:22:34 crc kubenswrapper[4862]: E1128 11:22:34.903355 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: E1128 11:22:34.903599 4862 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.903623 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.906608 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.906699 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.906718 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.906746 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.906765 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:34Z","lastTransitionTime":"2025-11-28T11:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.918243 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.937446 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.952543 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.968159 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:34Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.986620 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwcqx\" (UniqueName: \"kubernetes.io/projected/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-kube-api-access-wwcqx\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.986724 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.986801 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:34 crc kubenswrapper[4862]: I1128 11:22:34.986837 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.003441 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.009987 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.010183 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.010278 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.010351 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.010408 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.026730 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.049257 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.069662 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.087824 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.087908 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.087939 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.087966 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwcqx\" (UniqueName: \"kubernetes.io/projected/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-kube-api-access-wwcqx\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.089640 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.089879 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.092305 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.103161 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.114581 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.114642 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.114664 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.114693 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.114714 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.114913 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.117505 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwcqx\" (UniqueName: \"kubernetes.io/projected/4dfd45e3-0af1-469f-ab6a-4fd5ff606718-kube-api-access-wwcqx\") pod \"ovnkube-control-plane-749d76644c-lg2cn\" (UID: \"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.132914 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.153988 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.174304 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.204116 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:33Z\\\",\\\"message\\\":\\\"7973 6144 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1128 11:22:32.838039 6144 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 11:22:32.837333 6144 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:32.838567 6144 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:32.838589 6144 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:32.838611 6144 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 11:22:32.838634 6144 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 11:22:32.838649 6144 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 11:22:32.838647 6144 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 11:22:32.838672 6144 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 11:22:32.838679 6144 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 11:22:32.838724 6144 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 11:22:32.838735 6144 factory.go:656] Stopping watch factory\\\\nI1128 11:22:32.838802 6144 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.217658 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.217694 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.217707 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.217728 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.217743 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.218361 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.226504 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: W1128 11:22:35.243687 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dfd45e3_0af1_469f_ab6a_4fd5ff606718.slice/crio-73ed6bedf0878927c472ef4a9e30e45145d8ca552b12ca0f25f40fbaac1c647b WatchSource:0}: Error finding container 73ed6bedf0878927c472ef4a9e30e45145d8ca552b12ca0f25f40fbaac1c647b: Status 404 returned error can't find the container with id 73ed6bedf0878927c472ef4a9e30e45145d8ca552b12ca0f25f40fbaac1c647b Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.256400 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:35Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.320890 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.320937 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.320946 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.320965 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.320977 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.424141 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.424191 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.424204 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.424230 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.424246 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.526774 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.526824 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.526837 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.526858 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.526872 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.629812 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.629876 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.629891 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.629915 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.629938 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.733073 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.733133 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.733144 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.733164 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.733177 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.836434 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.836507 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.836529 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.836559 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.836579 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.837828 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.837924 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:35 crc kubenswrapper[4862]: E1128 11:22:35.837978 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.837831 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:35 crc kubenswrapper[4862]: E1128 11:22:35.838158 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:35 crc kubenswrapper[4862]: E1128 11:22:35.838606 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.943915 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.943962 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.943974 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.943994 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:35 crc kubenswrapper[4862]: I1128 11:22:35.944008 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:35Z","lastTransitionTime":"2025-11-28T11:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.046919 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.046988 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.047005 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.047030 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.047048 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.147536 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/1.log" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.148166 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/0.log" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.149226 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.149256 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.149269 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.149288 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.149302 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.150933 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660" exitCode=1 Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.151007 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.151064 4862 scope.go:117] "RemoveContainer" containerID="17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.151841 4862 scope.go:117] "RemoveContainer" containerID="0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660" Nov 28 11:22:36 crc kubenswrapper[4862]: E1128 11:22:36.152011 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.155709 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" event={"ID":"4dfd45e3-0af1-469f-ab6a-4fd5ff606718","Type":"ContainerStarted","Data":"fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.155746 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" event={"ID":"4dfd45e3-0af1-469f-ab6a-4fd5ff606718","Type":"ContainerStarted","Data":"c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.155757 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" event={"ID":"4dfd45e3-0af1-469f-ab6a-4fd5ff606718","Type":"ContainerStarted","Data":"73ed6bedf0878927c472ef4a9e30e45145d8ca552b12ca0f25f40fbaac1c647b"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.172795 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.191428 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.207126 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.223038 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.239834 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.255552 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.255608 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.255623 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.255644 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.255661 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.259131 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.273740 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.292439 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.310898 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.335038 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:33Z\\\",\\\"message\\\":\\\"7973 6144 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1128 11:22:32.838039 6144 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 11:22:32.837333 6144 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:32.838567 6144 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:32.838589 6144 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:32.838611 6144 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 11:22:32.838634 6144 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 11:22:32.838649 6144 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 11:22:32.838647 6144 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 11:22:32.838672 6144 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 11:22:32.838679 6144 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 11:22:32.838724 6144 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 11:22:32.838735 6144 factory.go:656] Stopping watch factory\\\\nI1128 11:22:32.838802 6144 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:35Z\\\",\\\"message\\\":\\\" 11:22:35.865405 6290 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 11:22:35.865437 6290 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 11:22:35.865499 6290 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.865552 6290 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:35.865777 6290 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.866830 6290 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.867509 6290 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:35.867547 6290 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:35.867584 6290 factory.go:656] Stopping watch factory\\\\nI1128 11:22:35.867608 6290 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11:22:35.867680 6290 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.359383 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.359432 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.359445 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.359463 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.359479 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.361228 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.383125 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.406274 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.430910 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.445579 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.463279 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.463330 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.463342 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.463361 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.463374 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.465035 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.486977 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.510126 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.528833 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.546736 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.566544 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.566629 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.566650 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.566683 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.566704 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.567287 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.583644 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.617528 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:33Z\\\",\\\"message\\\":\\\"7973 6144 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1128 11:22:32.838039 6144 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 11:22:32.837333 6144 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:32.838567 6144 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:32.838589 6144 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:32.838611 6144 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 11:22:32.838634 6144 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 11:22:32.838649 6144 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 11:22:32.838647 6144 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 11:22:32.838672 6144 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 11:22:32.838679 6144 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 11:22:32.838724 6144 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 11:22:32.838735 6144 factory.go:656] Stopping watch factory\\\\nI1128 11:22:32.838802 6144 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:35Z\\\",\\\"message\\\":\\\" 11:22:35.865405 6290 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 11:22:35.865437 6290 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 11:22:35.865499 6290 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.865552 6290 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:35.865777 6290 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.866830 6290 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.867509 6290 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:35.867547 6290 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:35.867584 6290 factory.go:656] Stopping watch factory\\\\nI1128 11:22:35.867608 6290 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11:22:35.867680 6290 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.640228 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.661770 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.669974 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.670038 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.670057 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.670085 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.670143 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.686832 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.704736 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.722670 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.741136 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.773205 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.773267 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.773284 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.773312 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.773332 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.776596 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.802451 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.820751 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.821817 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-wl4cj"] Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.822612 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:36 crc kubenswrapper[4862]: E1128 11:22:36.822716 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.842825 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.862206 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.885616 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.885688 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.885709 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.885736 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.885757 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.889461 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.910028 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.910175 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7tgf\" (UniqueName: \"kubernetes.io/projected/60f0df65-b253-4ea2-abda-dbc50f7f2eca-kube-api-access-w7tgf\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.912329 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.933517 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.957866 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.977851 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:36Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.991583 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.991675 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.991706 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.991744 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:36 crc kubenswrapper[4862]: I1128 11:22:36.991771 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:36Z","lastTransitionTime":"2025-11-28T11:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.003263 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.011276 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7tgf\" (UniqueName: \"kubernetes.io/projected/60f0df65-b253-4ea2-abda-dbc50f7f2eca-kube-api-access-w7tgf\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.011392 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.011548 4862 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.011662 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs podName:60f0df65-b253-4ea2-abda-dbc50f7f2eca nodeName:}" failed. No retries permitted until 2025-11-28 11:22:37.511632833 +0000 UTC m=+35.044146794 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs") pod "network-metrics-daemon-wl4cj" (UID: "60f0df65-b253-4ea2-abda-dbc50f7f2eca") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.024602 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.042412 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7tgf\" (UniqueName: \"kubernetes.io/projected/60f0df65-b253-4ea2-abda-dbc50f7f2eca-kube-api-access-w7tgf\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.050593 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.084498 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:33Z\\\",\\\"message\\\":\\\"7973 6144 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1128 11:22:32.838039 6144 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 11:22:32.837333 6144 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:32.838567 6144 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:32.838589 6144 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:32.838611 6144 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 11:22:32.838634 6144 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 11:22:32.838649 6144 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 11:22:32.838647 6144 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 11:22:32.838672 6144 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 11:22:32.838679 6144 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 11:22:32.838724 6144 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 11:22:32.838735 6144 factory.go:656] Stopping watch factory\\\\nI1128 11:22:32.838802 6144 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:35Z\\\",\\\"message\\\":\\\" 11:22:35.865405 6290 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 11:22:35.865437 6290 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 11:22:35.865499 6290 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.865552 6290 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:35.865777 6290 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.866830 6290 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.867509 6290 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:35.867547 6290 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:35.867584 6290 factory.go:656] Stopping watch factory\\\\nI1128 11:22:35.867608 6290 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11:22:35.867680 6290 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.095790 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.095934 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.096018 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.096042 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.096063 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:37Z","lastTransitionTime":"2025-11-28T11:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.119068 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.144594 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.163001 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/1.log" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.176823 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.198655 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.198696 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.198707 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.198724 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.198736 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:37Z","lastTransitionTime":"2025-11-28T11:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.200716 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.215043 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.228666 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:37Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.301406 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.301464 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.301481 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.301504 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.301521 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:37Z","lastTransitionTime":"2025-11-28T11:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.404010 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.404084 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.404131 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.404153 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.404169 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:37Z","lastTransitionTime":"2025-11-28T11:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.506690 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.506750 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.506767 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.506793 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.506811 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:37Z","lastTransitionTime":"2025-11-28T11:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.516461 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.516637 4862 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.516726 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs podName:60f0df65-b253-4ea2-abda-dbc50f7f2eca nodeName:}" failed. No retries permitted until 2025-11-28 11:22:38.51670308 +0000 UTC m=+36.049217041 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs") pod "network-metrics-daemon-wl4cj" (UID: "60f0df65-b253-4ea2-abda-dbc50f7f2eca") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.610434 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.610896 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.610916 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.610945 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.610963 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:37Z","lastTransitionTime":"2025-11-28T11:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.616965 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.617244 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:22:53.617209582 +0000 UTC m=+51.149723543 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.714252 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.714320 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.714337 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.714363 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.714380 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:37Z","lastTransitionTime":"2025-11-28T11:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.717929 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.717997 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.718045 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.718085 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718190 4862 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718233 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718271 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:53.718248721 +0000 UTC m=+51.250762752 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718274 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718278 4862 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718302 4862 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718354 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:53.718333603 +0000 UTC m=+51.250847564 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718389 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:53.718365504 +0000 UTC m=+51.250879455 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718449 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718516 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718533 4862 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.718595 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 11:22:53.71857943 +0000 UTC m=+51.251093391 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.817598 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.817691 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.817714 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.817747 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.817770 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:37Z","lastTransitionTime":"2025-11-28T11:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.838293 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.838322 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.838397 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.838630 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.838786 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:37 crc kubenswrapper[4862]: E1128 11:22:37.838946 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.920759 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.920851 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.920881 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.920912 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:37 crc kubenswrapper[4862]: I1128 11:22:37.920935 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:37Z","lastTransitionTime":"2025-11-28T11:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.023415 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.023485 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.023505 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.023586 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.023614 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.127257 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.127318 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.127335 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.127359 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.127376 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.230643 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.230704 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.230723 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.230745 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.230763 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.334486 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.334582 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.334596 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.334613 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.334629 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.438177 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.438247 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.438264 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.438289 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.438309 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.526693 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:38 crc kubenswrapper[4862]: E1128 11:22:38.526874 4862 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:38 crc kubenswrapper[4862]: E1128 11:22:38.526960 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs podName:60f0df65-b253-4ea2-abda-dbc50f7f2eca nodeName:}" failed. No retries permitted until 2025-11-28 11:22:40.526937187 +0000 UTC m=+38.059451138 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs") pod "network-metrics-daemon-wl4cj" (UID: "60f0df65-b253-4ea2-abda-dbc50f7f2eca") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.541203 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.541273 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.541312 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.541345 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.541368 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.645956 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.646054 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.646086 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.646176 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.646206 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.748887 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.748962 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.748986 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.749014 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.749037 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.837890 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:38 crc kubenswrapper[4862]: E1128 11:22:38.838365 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.852273 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.852352 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.852372 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.852396 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.852417 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.912620 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.949934 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:33Z\\\",\\\"message\\\":\\\"7973 6144 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1128 11:22:32.838039 6144 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 11:22:32.837333 6144 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:32.838567 6144 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:32.838589 6144 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:32.838611 6144 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 11:22:32.838634 6144 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 11:22:32.838649 6144 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 11:22:32.838647 6144 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 11:22:32.838672 6144 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 11:22:32.838679 6144 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 11:22:32.838724 6144 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 11:22:32.838735 6144 factory.go:656] Stopping watch factory\\\\nI1128 11:22:32.838802 6144 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:35Z\\\",\\\"message\\\":\\\" 11:22:35.865405 6290 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 11:22:35.865437 6290 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 11:22:35.865499 6290 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.865552 6290 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:35.865777 6290 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.866830 6290 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.867509 6290 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:35.867547 6290 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:35.867584 6290 factory.go:656] Stopping watch factory\\\\nI1128 11:22:35.867608 6290 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11:22:35.867680 6290 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:38Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.955713 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.955805 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.955826 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.955851 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.955869 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:38Z","lastTransitionTime":"2025-11-28T11:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.970673 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:38Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:38 crc kubenswrapper[4862]: I1128 11:22:38.989528 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:38Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.010817 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.030239 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.045220 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.058807 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.058855 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.058870 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.058891 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.058905 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.061778 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.093174 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.113656 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.130864 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.151759 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.161571 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.161618 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.161628 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.161645 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.161659 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.169790 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.190536 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.209237 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.226700 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.244789 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.264875 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.264925 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.264949 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.264974 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.264993 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.267392 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:39Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.368593 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.368645 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.368665 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.368688 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.368706 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.471742 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.471825 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.471845 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.471873 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.471898 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.575274 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.575364 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.575390 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.575425 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.575458 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.679594 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.679641 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.679654 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.679671 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.679685 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.782255 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.782318 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.782337 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.782361 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.782380 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.837949 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.837983 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:39 crc kubenswrapper[4862]: E1128 11:22:39.838230 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.837982 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:39 crc kubenswrapper[4862]: E1128 11:22:39.838400 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:39 crc kubenswrapper[4862]: E1128 11:22:39.838589 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.884571 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.884620 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.884628 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.884642 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.884651 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.987670 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.987747 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.987766 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.987792 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:39 crc kubenswrapper[4862]: I1128 11:22:39.987814 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:39Z","lastTransitionTime":"2025-11-28T11:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.090325 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.090397 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.090417 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.090440 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.090459 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:40Z","lastTransitionTime":"2025-11-28T11:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.193598 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.193699 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.193725 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.193765 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.193792 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:40Z","lastTransitionTime":"2025-11-28T11:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.297560 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.297633 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.297662 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.297693 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.297710 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:40Z","lastTransitionTime":"2025-11-28T11:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.400709 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.400778 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.400801 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.400829 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.400852 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:40Z","lastTransitionTime":"2025-11-28T11:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.504486 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.504549 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.504571 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.504594 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.504617 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:40Z","lastTransitionTime":"2025-11-28T11:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.550442 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:40 crc kubenswrapper[4862]: E1128 11:22:40.550692 4862 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:40 crc kubenswrapper[4862]: E1128 11:22:40.550873 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs podName:60f0df65-b253-4ea2-abda-dbc50f7f2eca nodeName:}" failed. No retries permitted until 2025-11-28 11:22:44.550840659 +0000 UTC m=+42.083354630 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs") pod "network-metrics-daemon-wl4cj" (UID: "60f0df65-b253-4ea2-abda-dbc50f7f2eca") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.607967 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.608059 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.608084 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.608147 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.608167 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:40Z","lastTransitionTime":"2025-11-28T11:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.710926 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.711012 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.711038 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.711068 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.711133 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:40Z","lastTransitionTime":"2025-11-28T11:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.813625 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.813679 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.813696 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.813719 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.813740 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:40Z","lastTransitionTime":"2025-11-28T11:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.838417 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:40 crc kubenswrapper[4862]: E1128 11:22:40.838618 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.916205 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.916274 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.916293 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.916318 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:40 crc kubenswrapper[4862]: I1128 11:22:40.916337 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:40Z","lastTransitionTime":"2025-11-28T11:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.019564 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.019600 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.019608 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.019624 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.019637 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.123336 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.123406 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.123426 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.123455 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.123474 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.226673 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.226712 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.226726 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.226743 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.226755 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.330257 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.330318 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.330339 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.330360 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.330377 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.433321 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.433367 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.433380 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.433398 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.433408 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.536383 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.536420 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.536431 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.536462 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.536475 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.639458 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.639505 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.639519 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.639537 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.639550 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.742768 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.742803 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.742812 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.742827 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.742889 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.837928 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.838047 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:41 crc kubenswrapper[4862]: E1128 11:22:41.838189 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:41 crc kubenswrapper[4862]: E1128 11:22:41.838267 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.837942 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:41 crc kubenswrapper[4862]: E1128 11:22:41.838383 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.852017 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.852335 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.852420 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.852461 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.852487 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.957223 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.957280 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.957295 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.957318 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:41 crc kubenswrapper[4862]: I1128 11:22:41.957335 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:41Z","lastTransitionTime":"2025-11-28T11:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.060161 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.060196 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.060206 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.060219 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.060228 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.163735 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.163774 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.163785 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.163802 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.163817 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.267275 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.267577 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.267628 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.267660 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.267685 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.370744 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.370790 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.370804 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.370823 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.370852 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.473675 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.473725 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.473737 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.473756 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.473767 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.576469 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.576540 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.576552 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.576575 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.576591 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.679269 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.679311 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.679328 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.679350 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.679368 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.780901 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.780987 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.781000 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.781016 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.781029 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.838146 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:42 crc kubenswrapper[4862]: E1128 11:22:42.838418 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.861058 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:42Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.883549 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:42Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.884791 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.884849 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.884869 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.884900 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.884923 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.906750 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:42Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.935980 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:42Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.956852 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:42Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.974702 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:42Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.987617 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.987662 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.987674 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.987690 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.987702 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:42Z","lastTransitionTime":"2025-11-28T11:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:42 crc kubenswrapper[4862]: I1128 11:22:42.990761 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:42Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.008015 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.027776 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.054602 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.085738 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b73ccb563f8e477c155cdfe3d1f79b9decb74133d457fc1481615db1427cd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:33Z\\\",\\\"message\\\":\\\"7973 6144 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1128 11:22:32.838039 6144 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 11:22:32.837333 6144 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:32.838567 6144 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:32.838589 6144 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:32.838611 6144 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 11:22:32.838634 6144 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 11:22:32.838649 6144 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 11:22:32.838647 6144 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 11:22:32.838672 6144 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 11:22:32.838679 6144 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 11:22:32.838724 6144 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 11:22:32.838735 6144 factory.go:656] Stopping watch factory\\\\nI1128 11:22:32.838802 6144 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:35Z\\\",\\\"message\\\":\\\" 11:22:35.865405 6290 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 11:22:35.865437 6290 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 11:22:35.865499 6290 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.865552 6290 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:35.865777 6290 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.866830 6290 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.867509 6290 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:35.867547 6290 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:35.867584 6290 factory.go:656] Stopping watch factory\\\\nI1128 11:22:35.867608 6290 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11:22:35.867680 6290 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.090331 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.090384 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.090398 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.090416 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.090431 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:43Z","lastTransitionTime":"2025-11-28T11:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.113807 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.130530 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.145649 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.164390 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.178719 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.192563 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.192600 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.192609 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.192621 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.192630 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:43Z","lastTransitionTime":"2025-11-28T11:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.196052 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:43Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.296525 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.296589 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.296606 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.296638 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.296665 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:43Z","lastTransitionTime":"2025-11-28T11:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.399674 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.399717 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.399728 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.399747 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.399759 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:43Z","lastTransitionTime":"2025-11-28T11:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.502250 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.502288 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.502314 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.502331 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.502341 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:43Z","lastTransitionTime":"2025-11-28T11:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.606219 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.606293 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.606312 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.606342 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.606366 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:43Z","lastTransitionTime":"2025-11-28T11:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.709588 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.709654 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.709675 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.709703 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.709724 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:43Z","lastTransitionTime":"2025-11-28T11:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.814068 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.814194 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.814215 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.814244 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.814267 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:43Z","lastTransitionTime":"2025-11-28T11:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.838408 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.838514 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:43 crc kubenswrapper[4862]: E1128 11:22:43.838574 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.838507 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:43 crc kubenswrapper[4862]: E1128 11:22:43.838720 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:43 crc kubenswrapper[4862]: E1128 11:22:43.838918 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.917996 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.918049 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.918061 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.918081 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:43 crc kubenswrapper[4862]: I1128 11:22:43.918402 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:43Z","lastTransitionTime":"2025-11-28T11:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.022995 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.023228 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.023300 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.023338 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.023403 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.127167 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.127250 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.127272 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.127302 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.127322 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.231451 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.231532 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.231551 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.231581 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.231602 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.335774 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.335902 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.335962 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.335991 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.336011 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.439775 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.439831 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.439842 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.439860 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.439873 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.542434 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.542512 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.542531 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.542560 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.542582 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.593408 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:44 crc kubenswrapper[4862]: E1128 11:22:44.593806 4862 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:44 crc kubenswrapper[4862]: E1128 11:22:44.593991 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs podName:60f0df65-b253-4ea2-abda-dbc50f7f2eca nodeName:}" failed. No retries permitted until 2025-11-28 11:22:52.59393165 +0000 UTC m=+50.126445611 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs") pod "network-metrics-daemon-wl4cj" (UID: "60f0df65-b253-4ea2-abda-dbc50f7f2eca") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.646557 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.646622 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.646639 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.646669 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.646693 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.750393 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.750476 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.750494 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.750524 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.750543 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.838941 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:44 crc kubenswrapper[4862]: E1128 11:22:44.839221 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.854332 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.854393 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.854429 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.854454 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.854474 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.958633 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.959005 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.959248 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.959404 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:44 crc kubenswrapper[4862]: I1128 11:22:44.959549 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:44Z","lastTransitionTime":"2025-11-28T11:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.062791 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.062830 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.062839 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.062854 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.062867 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.166125 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.166169 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.166183 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.166202 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.166214 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.263192 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.263255 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.263274 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.263298 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.263319 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: E1128 11:22:45.282935 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:45Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.289141 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.289189 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.289210 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.289233 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.289250 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: E1128 11:22:45.307872 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:45Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.312929 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.313004 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.313023 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.313045 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.313064 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: E1128 11:22:45.332246 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:45Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.337789 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.337840 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.337856 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.337881 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.337896 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: E1128 11:22:45.358477 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:45Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.364523 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.364591 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.364607 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.364632 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.364647 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: E1128 11:22:45.384613 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:45Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:45 crc kubenswrapper[4862]: E1128 11:22:45.384788 4862 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.387435 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.387513 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.387531 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.387556 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.387574 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.490209 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.490266 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.490284 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.490308 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.490326 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.593345 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.593424 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.593448 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.593479 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.593503 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.696171 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.696225 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.696241 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.696263 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.696279 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.799539 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.799596 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.799612 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.799635 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.799651 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.838287 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.838456 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.838461 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:45 crc kubenswrapper[4862]: E1128 11:22:45.838699 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:45 crc kubenswrapper[4862]: E1128 11:22:45.838884 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:45 crc kubenswrapper[4862]: E1128 11:22:45.839071 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.902448 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.902503 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.902519 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.902542 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:45 crc kubenswrapper[4862]: I1128 11:22:45.902559 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:45Z","lastTransitionTime":"2025-11-28T11:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.005799 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.005872 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.005897 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.005925 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.005944 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.108943 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.109013 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.109033 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.109057 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.109078 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.212817 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.212862 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.212878 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.212902 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.212920 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.316466 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.316529 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.316546 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.316603 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.316630 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.420770 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.420841 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.420861 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.420886 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.420904 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.523776 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.523860 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.523887 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.523918 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.523941 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.627199 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.627300 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.627323 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.627351 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.627371 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.734579 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.735326 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.735349 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.735383 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.735406 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.837787 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:46 crc kubenswrapper[4862]: E1128 11:22:46.837987 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.838074 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.838177 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.838199 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.838258 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.838287 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.941336 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.941390 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.941408 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.941431 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:46 crc kubenswrapper[4862]: I1128 11:22:46.941448 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:46Z","lastTransitionTime":"2025-11-28T11:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.044469 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.044712 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.044731 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.044753 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.044770 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.147445 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.147505 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.147521 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.147546 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.147563 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.250051 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.250148 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.250167 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.250198 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.250221 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.353735 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.353808 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.353830 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.353862 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.353887 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.457521 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.457605 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.457633 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.457661 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.457679 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.560712 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.560799 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.560817 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.560842 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.560860 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.569039 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.570294 4862 scope.go:117] "RemoveContainer" containerID="0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.588500 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.610366 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.638359 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.661565 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.663830 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.663878 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.663900 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.663929 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.663950 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.686765 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.705765 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.726168 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.749624 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.768420 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.769004 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.769016 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.769038 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.769052 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.770615 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.789787 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.822272 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:35Z\\\",\\\"message\\\":\\\" 11:22:35.865405 6290 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 11:22:35.865437 6290 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 11:22:35.865499 6290 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.865552 6290 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:35.865777 6290 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.866830 6290 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.867509 6290 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:35.867547 6290 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:35.867584 6290 factory.go:656] Stopping watch factory\\\\nI1128 11:22:35.867608 6290 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11:22:35.867680 6290 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.838019 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.838105 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.838041 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:47 crc kubenswrapper[4862]: E1128 11:22:47.838242 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:47 crc kubenswrapper[4862]: E1128 11:22:47.838346 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:47 crc kubenswrapper[4862]: E1128 11:22:47.838501 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.846982 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.864502 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.872945 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.872992 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.873009 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.873028 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.873043 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.882992 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.904130 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.921064 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.941042 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:47Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.976454 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.976524 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.976534 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.976554 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:47 crc kubenswrapper[4862]: I1128 11:22:47.976566 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:47Z","lastTransitionTime":"2025-11-28T11:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.080047 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.080084 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.080107 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.080122 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.080133 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:48Z","lastTransitionTime":"2025-11-28T11:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.183121 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.183178 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.183190 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.183209 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.183224 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:48Z","lastTransitionTime":"2025-11-28T11:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.233819 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/1.log" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.237477 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.238130 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.253934 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.269258 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.285542 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.285582 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.285592 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.285611 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.285622 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:48Z","lastTransitionTime":"2025-11-28T11:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.294466 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.313314 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.331456 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.351117 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.370904 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.381621 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.388716 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.388754 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.388766 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.388783 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.388798 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:48Z","lastTransitionTime":"2025-11-28T11:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.392532 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.404017 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.418685 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.434010 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.446681 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.459121 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.473719 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.490903 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.490962 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.490975 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.490998 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.491016 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:48Z","lastTransitionTime":"2025-11-28T11:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.497326 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.528607 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:35Z\\\",\\\"message\\\":\\\" 11:22:35.865405 6290 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 11:22:35.865437 6290 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 11:22:35.865499 6290 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.865552 6290 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:35.865777 6290 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.866830 6290 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.867509 6290 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:35.867547 6290 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:35.867584 6290 factory.go:656] Stopping watch factory\\\\nI1128 11:22:35.867608 6290 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11:22:35.867680 6290 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:48Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.593996 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.594082 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.594138 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.594171 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.594196 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:48Z","lastTransitionTime":"2025-11-28T11:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.696597 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.696672 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.696695 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.696728 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.696751 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:48Z","lastTransitionTime":"2025-11-28T11:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.805161 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.805221 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.805234 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.805261 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.805276 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:48Z","lastTransitionTime":"2025-11-28T11:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.838769 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:48 crc kubenswrapper[4862]: E1128 11:22:48.838995 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.908972 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.909038 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.909056 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.909086 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:48 crc kubenswrapper[4862]: I1128 11:22:48.909129 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:48Z","lastTransitionTime":"2025-11-28T11:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.012788 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.012837 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.012853 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.012874 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.012892 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.115849 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.115911 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.115930 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.115955 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.115972 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.219441 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.219517 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.219531 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.219549 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.219563 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.242340 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/2.log" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.243393 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/1.log" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.247506 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f" exitCode=1 Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.247562 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.247613 4862 scope.go:117] "RemoveContainer" containerID="0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.248694 4862 scope.go:117] "RemoveContainer" containerID="e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f" Nov 28 11:22:49 crc kubenswrapper[4862]: E1128 11:22:49.248943 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.271585 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.286510 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.297604 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.309524 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.322711 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.323081 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.323337 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.323512 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.323634 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.329322 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.344301 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.367629 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.386020 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.407413 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.427251 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.427317 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.427335 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.427361 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.427380 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.430329 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.448455 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.465755 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.484787 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.505778 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.530569 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.530639 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.530658 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.530684 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.530703 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.544033 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0580bbba50b27a4ec8b1b8e5bdeb2beb5ce4bf777be74d313aa708ee637be660\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:35Z\\\",\\\"message\\\":\\\" 11:22:35.865405 6290 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 11:22:35.865437 6290 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 11:22:35.865499 6290 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.865552 6290 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:35.865777 6290 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.866830 6290 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:35.867509 6290 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 11:22:35.867547 6290 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 11:22:35.867584 6290 factory.go:656] Stopping watch factory\\\\nI1128 11:22:35.867608 6290 ovnkube.go:599] Stopped ovnkube\\\\nI1128 11:22:35.867680 6290 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.569169 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.591270 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:49Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.634563 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.634630 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.634649 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.634677 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.634694 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.737926 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.737990 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.738010 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.738034 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.738056 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.838605 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.838706 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:49 crc kubenswrapper[4862]: E1128 11:22:49.838784 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.838605 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:49 crc kubenswrapper[4862]: E1128 11:22:49.838957 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:49 crc kubenswrapper[4862]: E1128 11:22:49.839254 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.840835 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.840904 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.840930 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.840956 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.840975 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.944010 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.944076 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.944129 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.944162 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:49 crc kubenswrapper[4862]: I1128 11:22:49.944187 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:49Z","lastTransitionTime":"2025-11-28T11:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.047380 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.047441 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.047461 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.047486 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.047507 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.150959 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.151016 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.151033 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.151058 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.151075 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.253055 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.253113 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.253125 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.253139 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.253148 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.253617 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/2.log" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.259401 4862 scope.go:117] "RemoveContainer" containerID="e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f" Nov 28 11:22:50 crc kubenswrapper[4862]: E1128 11:22:50.259732 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.283075 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.301647 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.318926 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.334074 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.355019 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.356748 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.356782 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.356852 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.356874 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.356886 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.376298 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.396891 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.416254 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.435390 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.464147 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.466020 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.466073 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.466120 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.466152 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.466171 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.498718 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.516189 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.534853 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.570461 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.570578 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.570597 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.570622 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.570640 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.571467 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.595373 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.615824 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.638365 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:50Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.673421 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.673503 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.673517 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.673542 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.673557 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.777660 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.777752 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.777778 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.777810 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.777835 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.838604 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:50 crc kubenswrapper[4862]: E1128 11:22:50.838856 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.880742 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.880811 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.880824 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.880844 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.880874 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.983927 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.983992 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.984016 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.984051 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:50 crc kubenswrapper[4862]: I1128 11:22:50.984073 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:50Z","lastTransitionTime":"2025-11-28T11:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.087598 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.087684 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.087707 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.087734 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.087756 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:51Z","lastTransitionTime":"2025-11-28T11:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.190872 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.190940 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.190956 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.190984 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.191004 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:51Z","lastTransitionTime":"2025-11-28T11:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.293622 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.293708 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.293725 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.293751 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.293770 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:51Z","lastTransitionTime":"2025-11-28T11:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.397506 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.397577 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.397596 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.397620 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.397639 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:51Z","lastTransitionTime":"2025-11-28T11:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.500433 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.500483 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.500501 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.500524 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.500541 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:51Z","lastTransitionTime":"2025-11-28T11:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.602978 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.603046 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.603063 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.603087 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.603128 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:51Z","lastTransitionTime":"2025-11-28T11:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.706261 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.706324 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.706346 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.706369 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.706382 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:51Z","lastTransitionTime":"2025-11-28T11:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.809744 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.809803 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.809822 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.809844 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.809862 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:51Z","lastTransitionTime":"2025-11-28T11:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.838119 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:51 crc kubenswrapper[4862]: E1128 11:22:51.838285 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.838147 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.838119 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:51 crc kubenswrapper[4862]: E1128 11:22:51.838378 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:51 crc kubenswrapper[4862]: E1128 11:22:51.838593 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.912264 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.912342 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.912351 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.912376 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:51 crc kubenswrapper[4862]: I1128 11:22:51.912389 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:51Z","lastTransitionTime":"2025-11-28T11:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.015047 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.015109 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.015124 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.015143 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.015155 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.118011 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.118088 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.118133 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.118159 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.118179 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.221696 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.221761 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.221778 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.221803 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.221820 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.325522 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.325610 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.325638 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.325672 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.325697 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.428892 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.428971 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.428993 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.429023 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.429041 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.532008 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.532070 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.532127 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.532155 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.532172 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.601347 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:52 crc kubenswrapper[4862]: E1128 11:22:52.601612 4862 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:52 crc kubenswrapper[4862]: E1128 11:22:52.601730 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs podName:60f0df65-b253-4ea2-abda-dbc50f7f2eca nodeName:}" failed. No retries permitted until 2025-11-28 11:23:08.601699448 +0000 UTC m=+66.134213399 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs") pod "network-metrics-daemon-wl4cj" (UID: "60f0df65-b253-4ea2-abda-dbc50f7f2eca") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.635263 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.635331 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.635357 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.635386 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.635409 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.737637 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.737694 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.737712 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.737736 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.737753 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.838627 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:52 crc kubenswrapper[4862]: E1128 11:22:52.839333 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.842381 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.842464 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.842488 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.842519 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.842540 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.856558 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:52Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.882166 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:52Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.903422 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:52Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.928476 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:52Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.945678 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.945771 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.945790 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.945816 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.945870 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:52Z","lastTransitionTime":"2025-11-28T11:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.950924 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:52Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.962733 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:52Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.977455 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:52Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:52 crc kubenswrapper[4862]: I1128 11:22:52.990935 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:52Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.002669 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:53Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.027517 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:53Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.049323 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.049388 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.049405 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.049434 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.049453 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.063005 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:53Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.091303 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:53Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.109336 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:53Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.125237 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:53Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.140708 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:53Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.151836 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.151895 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.151915 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.151940 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.151957 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.155039 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:53Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.166903 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:53Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.256346 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.256422 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.256447 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.256478 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.256501 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.359229 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.359302 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.359341 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.359374 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.359396 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.462113 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.462164 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.462176 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.462194 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.462205 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.564905 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.564973 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.564991 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.565016 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.565036 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.668614 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.668697 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.668724 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.668755 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.668791 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.713015 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.713259 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:23:25.71321763 +0000 UTC m=+83.245731591 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.771674 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.771744 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.771763 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.771789 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.771808 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.815237 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.815315 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.815387 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.815428 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815560 4862 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815593 4862 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815649 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:23:25.815626577 +0000 UTC m=+83.348140528 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815654 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815736 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815762 4862 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815672 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815947 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815984 4862 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.815695 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:23:25.815665938 +0000 UTC m=+83.348179899 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.816048 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 11:23:25.816031249 +0000 UTC m=+83.348545210 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.816082 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 11:23:25.81606688 +0000 UTC m=+83.348580831 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.838442 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.838561 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.838463 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.838718 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.838908 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:53 crc kubenswrapper[4862]: E1128 11:22:53.839022 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.874946 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.875027 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.875047 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.875072 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.875131 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.977759 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.977807 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.977820 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.977839 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:53 crc kubenswrapper[4862]: I1128 11:22:53.977854 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:53Z","lastTransitionTime":"2025-11-28T11:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.081318 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.081420 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.081439 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.081471 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.081497 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:54Z","lastTransitionTime":"2025-11-28T11:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.184833 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.184910 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.184933 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.184964 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.184989 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:54Z","lastTransitionTime":"2025-11-28T11:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.287940 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.288001 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.288019 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.288044 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.288064 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:54Z","lastTransitionTime":"2025-11-28T11:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.391388 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.391460 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.391479 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.391505 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.391523 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:54Z","lastTransitionTime":"2025-11-28T11:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.494172 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.494706 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.494724 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.494748 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.494770 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:54Z","lastTransitionTime":"2025-11-28T11:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.598530 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.598588 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.598608 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.598634 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.598651 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:54Z","lastTransitionTime":"2025-11-28T11:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.701489 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.701544 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.701560 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.701585 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.701603 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:54Z","lastTransitionTime":"2025-11-28T11:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.804240 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.804337 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.804355 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.804380 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.804397 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:54Z","lastTransitionTime":"2025-11-28T11:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.838434 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:54 crc kubenswrapper[4862]: E1128 11:22:54.838657 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.907453 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.907518 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.907533 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.907556 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:54 crc kubenswrapper[4862]: I1128 11:22:54.907570 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:54Z","lastTransitionTime":"2025-11-28T11:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.011315 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.011373 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.011394 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.011418 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.011436 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.114910 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.115277 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.115468 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.115621 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.115742 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.218575 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.218644 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.218661 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.218690 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.218710 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.329294 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.329392 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.329413 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.329444 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.329467 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.433596 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.434649 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.434890 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.435142 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.435324 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.539451 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.539527 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.539545 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.539580 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.539606 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.642605 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.642662 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.642678 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.642702 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.642725 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.708422 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.708505 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.708523 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.708552 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.708571 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: E1128 11:22:55.734848 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:55Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.741460 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.741590 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.741611 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.741637 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.741654 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: E1128 11:22:55.769612 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:55Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.775290 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.775358 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.775434 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.775512 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.775554 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: E1128 11:22:55.797217 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:55Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.803476 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.803528 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.803547 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.803568 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.803584 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: E1128 11:22:55.824010 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:55Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.830441 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.830528 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.830547 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.830568 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.830586 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.837886 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.837934 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:55 crc kubenswrapper[4862]: E1128 11:22:55.838008 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.838076 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:55 crc kubenswrapper[4862]: E1128 11:22:55.838161 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:55 crc kubenswrapper[4862]: E1128 11:22:55.838319 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:55 crc kubenswrapper[4862]: E1128 11:22:55.851851 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:55Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:55 crc kubenswrapper[4862]: E1128 11:22:55.852228 4862 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.854852 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.854952 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.854978 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.855012 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.855037 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.957981 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.958038 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.958055 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.958079 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:55 crc kubenswrapper[4862]: I1128 11:22:55.958125 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:55Z","lastTransitionTime":"2025-11-28T11:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.061052 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.061195 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.061228 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.061265 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.061289 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.163894 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.164468 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.164602 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.164758 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.164900 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.182447 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.201399 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.203298 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.225524 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.246997 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.262833 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.268692 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.268737 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.268754 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.268778 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.268798 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.282957 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.300900 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.324241 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.346235 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.372762 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.372840 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.372862 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.372889 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.372908 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.387627 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.411072 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.430811 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.449335 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.470671 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.476335 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.476394 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.476412 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.476441 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.476459 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.488154 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.507198 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.541121 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.561408 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:22:56Z is after 2025-08-24T17:21:41Z" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.580273 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.580539 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.580687 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.580834 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.580993 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.685853 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.685940 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.685965 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.685993 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.686011 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.789569 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.789655 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.789683 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.789716 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.789742 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.838884 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:56 crc kubenswrapper[4862]: E1128 11:22:56.839262 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.892841 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.892905 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.892924 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.892956 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.892978 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.996079 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.996215 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.996237 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.996268 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:56 crc kubenswrapper[4862]: I1128 11:22:56.996294 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:56Z","lastTransitionTime":"2025-11-28T11:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.099610 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.099681 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.099698 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.099726 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.099743 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:57Z","lastTransitionTime":"2025-11-28T11:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.202463 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.202505 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.202514 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.202528 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.202537 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:57Z","lastTransitionTime":"2025-11-28T11:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.305146 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.305194 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.305203 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.305223 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.305235 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:57Z","lastTransitionTime":"2025-11-28T11:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.408937 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.409018 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.409041 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.409127 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.409158 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:57Z","lastTransitionTime":"2025-11-28T11:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.512781 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.512847 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.512872 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.512902 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.512924 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:57Z","lastTransitionTime":"2025-11-28T11:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.615501 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.615585 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.615603 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.615629 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.615647 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:57Z","lastTransitionTime":"2025-11-28T11:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.718977 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.719024 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.719040 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.719063 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.719080 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:57Z","lastTransitionTime":"2025-11-28T11:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.823280 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.823470 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.823494 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.823569 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.823599 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:57Z","lastTransitionTime":"2025-11-28T11:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.837757 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:57 crc kubenswrapper[4862]: E1128 11:22:57.838020 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.838345 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.838420 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:57 crc kubenswrapper[4862]: E1128 11:22:57.838598 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:57 crc kubenswrapper[4862]: E1128 11:22:57.838712 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.926635 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.926697 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.926714 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.926740 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:57 crc kubenswrapper[4862]: I1128 11:22:57.926759 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:57Z","lastTransitionTime":"2025-11-28T11:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.030180 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.030342 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.030364 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.030819 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.031014 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.134579 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.134796 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.134818 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.134847 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.134876 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.237974 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.238034 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.238044 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.238063 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.238075 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.341521 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.341601 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.341620 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.341648 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.341672 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.444840 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.444896 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.444913 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.444936 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.444953 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.547785 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.547840 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.547857 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.547879 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.547897 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.650812 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.650900 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.650925 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.650955 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.650974 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.754051 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.754133 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.754150 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.754173 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.754194 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.838063 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:22:58 crc kubenswrapper[4862]: E1128 11:22:58.838293 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.856571 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.856642 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.856657 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.856677 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.856692 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.960476 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.960557 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.960577 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.960607 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:58 crc kubenswrapper[4862]: I1128 11:22:58.960625 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:58Z","lastTransitionTime":"2025-11-28T11:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.064642 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.064716 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.064740 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.064775 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.064798 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.168162 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.168227 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.168245 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.168270 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.168286 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.271669 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.271729 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.271755 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.271787 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.271808 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.376011 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.376087 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.376137 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.376165 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.376191 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.479437 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.479493 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.479514 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.479541 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.479565 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.583293 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.583493 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.583582 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.583615 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.583639 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.686733 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.686783 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.686805 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.686827 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.686843 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.790663 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.790731 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.790984 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.791025 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.791047 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.838026 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.838074 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.838026 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:22:59 crc kubenswrapper[4862]: E1128 11:22:59.838249 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:22:59 crc kubenswrapper[4862]: E1128 11:22:59.838352 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:22:59 crc kubenswrapper[4862]: E1128 11:22:59.838467 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.893976 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.894114 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.894134 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.894156 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.894172 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.997584 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.997634 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.997650 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.997674 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:22:59 crc kubenswrapper[4862]: I1128 11:22:59.997691 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:22:59Z","lastTransitionTime":"2025-11-28T11:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.100216 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.100289 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.100312 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.100339 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.100361 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:00Z","lastTransitionTime":"2025-11-28T11:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.204132 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.204200 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.204222 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.204251 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.204273 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:00Z","lastTransitionTime":"2025-11-28T11:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.306908 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.306990 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.307015 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.307132 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.307152 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:00Z","lastTransitionTime":"2025-11-28T11:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.409984 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.410048 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.410076 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.410134 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.410151 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:00Z","lastTransitionTime":"2025-11-28T11:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.514564 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.514650 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.514672 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.514701 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.514731 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:00Z","lastTransitionTime":"2025-11-28T11:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.617764 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.617861 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.617885 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.617916 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.617943 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:00Z","lastTransitionTime":"2025-11-28T11:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.721005 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.721128 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.721158 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.721190 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.721214 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:00Z","lastTransitionTime":"2025-11-28T11:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.823785 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.823885 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.823939 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.823969 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.823987 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:00Z","lastTransitionTime":"2025-11-28T11:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.838656 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:00 crc kubenswrapper[4862]: E1128 11:23:00.838899 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.927579 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.927654 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.927676 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.927704 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:00 crc kubenswrapper[4862]: I1128 11:23:00.927727 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:00Z","lastTransitionTime":"2025-11-28T11:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.030709 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.030782 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.030805 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.030838 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.030861 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.134281 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.134369 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.134401 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.134434 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.134459 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.237936 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.237994 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.238016 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.238039 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.238057 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.341036 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.341141 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.341183 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.341234 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.341258 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.444921 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.444979 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.444996 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.445020 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.445039 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.548817 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.548880 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.548891 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.548912 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.548928 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.652026 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.652083 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.652116 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.652139 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.652159 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.756124 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.756207 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.756230 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.756263 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.756291 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.838577 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.838651 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.838577 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:01 crc kubenswrapper[4862]: E1128 11:23:01.838779 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:01 crc kubenswrapper[4862]: E1128 11:23:01.838921 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:01 crc kubenswrapper[4862]: E1128 11:23:01.839040 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.840054 4862 scope.go:117] "RemoveContainer" containerID="e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f" Nov 28 11:23:01 crc kubenswrapper[4862]: E1128 11:23:01.840386 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.858925 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.858974 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.858992 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.859012 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.859030 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.962288 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.962333 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.962350 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.962378 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:01 crc kubenswrapper[4862]: I1128 11:23:01.962405 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:01Z","lastTransitionTime":"2025-11-28T11:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.066133 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.066188 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.066202 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.066222 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.066237 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.169880 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.169947 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.169963 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.169991 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.170011 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.273453 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.273516 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.273534 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.273559 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.273576 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.376471 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.376548 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.376559 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.376634 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.376651 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.479473 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.479554 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.479564 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.479586 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.479597 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.582762 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.582827 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.582844 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.582871 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.582889 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.686924 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.686994 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.687012 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.687037 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.687052 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.790007 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.790138 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.790157 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.790182 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.790230 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.837684 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:02 crc kubenswrapper[4862]: E1128 11:23:02.837866 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.880388 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:02Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.893832 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.893918 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.893936 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.893986 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.894004 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.901917 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:02Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.925832 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:02Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.961506 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:02Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.994915 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:02Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.996378 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.996419 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.996433 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.996454 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:02 crc kubenswrapper[4862]: I1128 11:23:02.996470 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:02Z","lastTransitionTime":"2025-11-28T11:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.017462 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.038727 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.058046 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.071371 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.089123 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.099799 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.099857 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.099868 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.099892 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.099906 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:03Z","lastTransitionTime":"2025-11-28T11:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.104178 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.121378 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ac1115e-d787-4ff7-b7f9-399137d327bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ca7d0113725fb7adcdea09bd4592580b9652ce2bb40e095484ec0c458203eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://069da5c66373778aec270d3597c665d396a164a6c07f258553405d7b269f3f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87d5fb88089c845017e6226c86fd1a41fc3622ad32b9ffec123ebb3649be4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.136521 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.161478 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.181361 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.202211 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.204229 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.204322 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.204358 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.204412 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.204430 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:03Z","lastTransitionTime":"2025-11-28T11:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.220813 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.236012 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:03Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.306747 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.306792 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.306809 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.306833 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.306851 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:03Z","lastTransitionTime":"2025-11-28T11:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.409982 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.410051 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.410070 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.410125 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.410142 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:03Z","lastTransitionTime":"2025-11-28T11:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.514037 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.514145 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.514171 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.514200 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.514221 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:03Z","lastTransitionTime":"2025-11-28T11:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.617063 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.617183 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.617210 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.617246 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.617271 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:03Z","lastTransitionTime":"2025-11-28T11:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.720011 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.721185 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.721330 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.721496 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.721637 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:03Z","lastTransitionTime":"2025-11-28T11:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.824943 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.825027 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.825051 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.825137 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.825168 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:03Z","lastTransitionTime":"2025-11-28T11:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.838376 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.838548 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.838401 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:03 crc kubenswrapper[4862]: E1128 11:23:03.838811 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:03 crc kubenswrapper[4862]: E1128 11:23:03.839147 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:03 crc kubenswrapper[4862]: E1128 11:23:03.839297 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.928916 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.928975 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.928992 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.929015 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:03 crc kubenswrapper[4862]: I1128 11:23:03.929033 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:03Z","lastTransitionTime":"2025-11-28T11:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.032189 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.032243 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.032261 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.032284 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.032300 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.135152 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.135203 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.135219 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.135244 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.135260 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.239267 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.239319 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.239336 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.239358 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.239377 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.342541 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.342605 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.342625 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.342648 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.342667 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.445675 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.445879 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.445901 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.445924 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.445942 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.550050 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.550160 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.550179 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.550202 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.550220 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.653645 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.653710 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.653729 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.653753 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.653770 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.756566 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.756635 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.756653 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.756681 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.756699 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.838550 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:04 crc kubenswrapper[4862]: E1128 11:23:04.838733 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.860171 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.860269 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.860320 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.860345 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.860367 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.964064 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.964163 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.964190 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.964216 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:04 crc kubenswrapper[4862]: I1128 11:23:04.964236 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:04Z","lastTransitionTime":"2025-11-28T11:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.067911 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.068044 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.068067 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.068148 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.068169 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.171256 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.171320 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.171337 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.171361 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.171379 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.273806 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.273876 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.273893 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.273915 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.273938 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.376551 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.376637 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.376661 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.376687 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.376709 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.480203 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.480285 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.480317 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.480349 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.480369 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.585302 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.585380 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.585405 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.585437 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.585460 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.688043 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.688149 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.688198 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.688232 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.688256 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.791808 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.791890 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.791911 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.791936 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.791953 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.838453 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.838538 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.838564 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:05 crc kubenswrapper[4862]: E1128 11:23:05.838693 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:05 crc kubenswrapper[4862]: E1128 11:23:05.838820 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:05 crc kubenswrapper[4862]: E1128 11:23:05.838984 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.895255 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.895379 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.895403 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.895429 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.895448 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.998734 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.998798 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.998815 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.998839 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:05 crc kubenswrapper[4862]: I1128 11:23:05.998858 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:05Z","lastTransitionTime":"2025-11-28T11:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.102218 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.102281 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.102299 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.102324 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.102340 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.204842 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.204999 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.205022 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.205052 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.205078 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.248554 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.248878 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.249312 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.249502 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.249572 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: E1128 11:23:06.269590 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:06Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.275211 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.275286 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.275306 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.275334 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.275353 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: E1128 11:23:06.298127 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:06Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.303767 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.303869 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.303897 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.303932 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.303963 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: E1128 11:23:06.327622 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:06Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.346792 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.346842 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.346855 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.346872 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.346884 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: E1128 11:23:06.367118 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:06Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.371679 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.371706 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.371716 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.371731 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.371740 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: E1128 11:23:06.387979 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:06Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:06 crc kubenswrapper[4862]: E1128 11:23:06.388113 4862 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.389502 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.389528 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.389539 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.389553 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.389563 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.491990 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.492019 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.492029 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.492042 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.492052 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.594914 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.594963 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.594989 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.595024 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.595048 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.698060 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.698131 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.698148 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.698170 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.698189 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.800363 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.800466 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.800484 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.800509 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.800525 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.837789 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:06 crc kubenswrapper[4862]: E1128 11:23:06.838023 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.903875 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.903934 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.903951 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.903977 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:06 crc kubenswrapper[4862]: I1128 11:23:06.903998 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:06Z","lastTransitionTime":"2025-11-28T11:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.007416 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.007473 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.007491 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.007516 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.007536 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.110301 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.110344 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.110355 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.110373 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.110388 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.213561 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.213633 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.213647 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.213669 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.213685 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.321689 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.322109 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.322122 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.322138 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.322151 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.424574 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.424649 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.424668 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.424693 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.424711 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.526669 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.526720 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.526734 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.526754 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.526767 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.631961 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.632000 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.632011 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.632026 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.632037 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.734711 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.734767 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.734786 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.734814 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.734831 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.837674 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.837722 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.837732 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.837749 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.837761 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.837783 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.837840 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:07 crc kubenswrapper[4862]: E1128 11:23:07.837872 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.837783 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:07 crc kubenswrapper[4862]: E1128 11:23:07.837992 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:07 crc kubenswrapper[4862]: E1128 11:23:07.838108 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.940435 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.940500 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.940518 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.940543 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:07 crc kubenswrapper[4862]: I1128 11:23:07.940560 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:07Z","lastTransitionTime":"2025-11-28T11:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.043440 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.043483 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.043494 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.043510 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.043521 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.145683 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.145729 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.145741 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.145757 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.145770 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.247859 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.247919 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.247937 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.247963 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.247981 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.350644 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.350692 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.350707 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.350725 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.350740 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.453921 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.453973 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.453984 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.454002 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.454017 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.556756 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.556814 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.556832 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.556855 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.556874 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.659486 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.659529 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.659542 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.659558 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.659569 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.696558 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:08 crc kubenswrapper[4862]: E1128 11:23:08.696811 4862 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:23:08 crc kubenswrapper[4862]: E1128 11:23:08.696911 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs podName:60f0df65-b253-4ea2-abda-dbc50f7f2eca nodeName:}" failed. No retries permitted until 2025-11-28 11:23:40.696888559 +0000 UTC m=+98.229402570 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs") pod "network-metrics-daemon-wl4cj" (UID: "60f0df65-b253-4ea2-abda-dbc50f7f2eca") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.762011 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.762070 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.762086 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.762134 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.762151 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.838610 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:08 crc kubenswrapper[4862]: E1128 11:23:08.838812 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.863750 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.863782 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.863792 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.863805 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.863815 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.965979 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.966027 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.966041 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.966059 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:08 crc kubenswrapper[4862]: I1128 11:23:08.966071 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:08Z","lastTransitionTime":"2025-11-28T11:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.068465 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.068533 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.068550 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.068575 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.068591 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.171484 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.171517 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.171525 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.171537 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.171547 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.274555 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.274603 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.274616 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.274632 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.274645 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.377713 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.377772 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.377782 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.377795 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.377806 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.480468 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.480527 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.480545 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.480568 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.480585 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.582669 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.582714 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.582727 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.582746 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.582774 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.685253 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.685310 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.685328 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.685352 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.685370 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.787861 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.787935 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.787960 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.787989 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.788010 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.837876 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.837952 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.837972 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:09 crc kubenswrapper[4862]: E1128 11:23:09.838136 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:09 crc kubenswrapper[4862]: E1128 11:23:09.838310 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:09 crc kubenswrapper[4862]: E1128 11:23:09.838460 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.891049 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.891078 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.891128 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.891157 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.891181 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.993711 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.993741 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.993750 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.993763 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:09 crc kubenswrapper[4862]: I1128 11:23:09.993771 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:09Z","lastTransitionTime":"2025-11-28T11:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.095858 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.095927 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.095951 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.095983 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.096007 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:10Z","lastTransitionTime":"2025-11-28T11:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.197925 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.197961 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.197971 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.197986 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.197996 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:10Z","lastTransitionTime":"2025-11-28T11:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.300492 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.300553 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.300569 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.300597 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.300614 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:10Z","lastTransitionTime":"2025-11-28T11:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.332927 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kqpd4_d6abffd5-6086-4a32-ba8d-e7c669548c5f/kube-multus/0.log" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.332984 4862 generic.go:334] "Generic (PLEG): container finished" podID="d6abffd5-6086-4a32-ba8d-e7c669548c5f" containerID="3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4" exitCode=1 Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.333016 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kqpd4" event={"ID":"d6abffd5-6086-4a32-ba8d-e7c669548c5f","Type":"ContainerDied","Data":"3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.333473 4862 scope.go:117] "RemoveContainer" containerID="3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.351335 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.367653 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.403393 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.403427 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.403436 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.403449 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.403459 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:10Z","lastTransitionTime":"2025-11-28T11:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.406245 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.424341 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.440387 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.456819 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:09Z\\\",\\\"message\\\":\\\"2025-11-28T11:22:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956\\\\n2025-11-28T11:22:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956 to /host/opt/cni/bin/\\\\n2025-11-28T11:22:24Z [verbose] multus-daemon started\\\\n2025-11-28T11:22:24Z [verbose] Readiness Indicator file check\\\\n2025-11-28T11:23:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.477318 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.495368 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.506361 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.506404 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.506413 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.506430 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.506440 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:10Z","lastTransitionTime":"2025-11-28T11:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.511202 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.526429 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.541077 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ac1115e-d787-4ff7-b7f9-399137d327bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ca7d0113725fb7adcdea09bd4592580b9652ce2bb40e095484ec0c458203eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://069da5c66373778aec270d3597c665d396a164a6c07f258553405d7b269f3f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87d5fb88089c845017e6226c86fd1a41fc3622ad32b9ffec123ebb3649be4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.560781 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.579712 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.598598 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.608987 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.609010 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.609021 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.609036 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.609046 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:10Z","lastTransitionTime":"2025-11-28T11:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.614817 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.631124 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.652259 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.672649 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:10Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.711837 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.711879 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.711889 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.711904 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.711913 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:10Z","lastTransitionTime":"2025-11-28T11:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.814316 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.814391 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.814416 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.814445 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.814470 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:10Z","lastTransitionTime":"2025-11-28T11:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.837703 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:10 crc kubenswrapper[4862]: E1128 11:23:10.838240 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.917026 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.917086 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.917135 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.917157 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:10 crc kubenswrapper[4862]: I1128 11:23:10.917175 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:10Z","lastTransitionTime":"2025-11-28T11:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.021582 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.021635 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.021656 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.021830 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.021857 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.125171 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.125244 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.125260 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.125284 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.125326 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.228628 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.228741 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.228761 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.228783 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.228800 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.331816 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.331874 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.331891 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.331921 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.331936 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.339914 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kqpd4_d6abffd5-6086-4a32-ba8d-e7c669548c5f/kube-multus/0.log" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.340014 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kqpd4" event={"ID":"d6abffd5-6086-4a32-ba8d-e7c669548c5f","Type":"ContainerStarted","Data":"743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.373902 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.398577 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.413940 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.432725 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:09Z\\\",\\\"message\\\":\\\"2025-11-28T11:22:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956\\\\n2025-11-28T11:22:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956 to /host/opt/cni/bin/\\\\n2025-11-28T11:22:24Z [verbose] multus-daemon started\\\\n2025-11-28T11:22:24Z [verbose] Readiness Indicator file check\\\\n2025-11-28T11:23:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.433990 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.434058 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.434078 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.434251 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.434288 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.448643 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.464109 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.476450 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.489421 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ac1115e-d787-4ff7-b7f9-399137d327bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ca7d0113725fb7adcdea09bd4592580b9652ce2bb40e095484ec0c458203eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://069da5c66373778aec270d3597c665d396a164a6c07f258553405d7b269f3f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87d5fb88089c845017e6226c86fd1a41fc3622ad32b9ffec123ebb3649be4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.506844 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.524060 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.537673 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.537741 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.537764 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.537796 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.537818 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.538319 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.561780 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.578166 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.592419 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.610450 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.626597 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.640603 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.640801 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.641053 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.641070 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.641082 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.648379 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.669540 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:11Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.743926 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.743997 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.744009 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.744027 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.744040 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.838068 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.838152 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.838203 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:11 crc kubenswrapper[4862]: E1128 11:23:11.838270 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:11 crc kubenswrapper[4862]: E1128 11:23:11.838402 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:11 crc kubenswrapper[4862]: E1128 11:23:11.838540 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.846954 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.846984 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.846993 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.847007 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.847015 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.949516 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.949561 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.949573 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.949588 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:11 crc kubenswrapper[4862]: I1128 11:23:11.949598 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:11Z","lastTransitionTime":"2025-11-28T11:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.052654 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.052695 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.052704 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.052721 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.052730 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.154856 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.154917 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.154935 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.154960 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.154978 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.258471 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.258713 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.258723 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.258741 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.258752 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.361477 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.361537 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.361550 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.361570 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.361587 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.465150 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.465204 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.465216 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.465233 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.465244 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.568432 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.568505 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.568523 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.568551 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.568571 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.671834 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.671904 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.671925 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.671954 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.671976 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.775276 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.775717 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.775886 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.777688 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.778150 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.837829 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:12 crc kubenswrapper[4862]: E1128 11:23:12.838047 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.856020 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:12Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.876674 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:12Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.881420 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.881473 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.881484 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.881500 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.881511 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.892129 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:09Z\\\",\\\"message\\\":\\\"2025-11-28T11:22:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956\\\\n2025-11-28T11:22:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956 to /host/opt/cni/bin/\\\\n2025-11-28T11:22:24Z [verbose] multus-daemon started\\\\n2025-11-28T11:22:24Z [verbose] Readiness Indicator file check\\\\n2025-11-28T11:23:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:12Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.903821 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:12Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.943784 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:12Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.971082 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:12Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.985627 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.985735 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.985816 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.985858 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.985882 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:12Z","lastTransitionTime":"2025-11-28T11:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:12 crc kubenswrapper[4862]: I1128 11:23:12.990240 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:12Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.004405 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.022065 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.038576 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.052961 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.068929 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.084936 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.089561 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.089614 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.089624 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.089641 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.089653 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:13Z","lastTransitionTime":"2025-11-28T11:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.106837 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ac1115e-d787-4ff7-b7f9-399137d327bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ca7d0113725fb7adcdea09bd4592580b9652ce2bb40e095484ec0c458203eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://069da5c66373778aec270d3597c665d396a164a6c07f258553405d7b269f3f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87d5fb88089c845017e6226c86fd1a41fc3622ad32b9ffec123ebb3649be4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.125903 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.146889 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.188371 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.192789 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.192862 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.192889 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.192918 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.192939 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:13Z","lastTransitionTime":"2025-11-28T11:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.206913 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:13Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.296355 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.296430 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.296455 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.296487 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.296509 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:13Z","lastTransitionTime":"2025-11-28T11:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.400009 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.400065 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.400082 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.400130 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.400169 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:13Z","lastTransitionTime":"2025-11-28T11:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.503163 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.503229 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.503252 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.503282 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.503306 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:13Z","lastTransitionTime":"2025-11-28T11:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.606217 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.606272 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.606291 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.606314 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.606330 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:13Z","lastTransitionTime":"2025-11-28T11:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.710298 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.710645 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.710974 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.711195 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.711405 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:13Z","lastTransitionTime":"2025-11-28T11:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.815000 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.815444 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.815634 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.815839 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.816032 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:13Z","lastTransitionTime":"2025-11-28T11:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.838467 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.838753 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:13 crc kubenswrapper[4862]: E1128 11:23:13.838942 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.838976 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:13 crc kubenswrapper[4862]: E1128 11:23:13.839312 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:13 crc kubenswrapper[4862]: E1128 11:23:13.839428 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.840585 4862 scope.go:117] "RemoveContainer" containerID="e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.919141 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.919649 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.919861 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.920049 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:13 crc kubenswrapper[4862]: I1128 11:23:13.920263 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:13Z","lastTransitionTime":"2025-11-28T11:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.026005 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.026622 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.026636 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.026655 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.026669 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.129472 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.129590 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.129643 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.129687 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.129706 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.232408 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.232453 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.232464 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.232481 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.232493 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.334768 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.334796 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.334804 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.334818 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.334829 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.352313 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/2.log" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.354615 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.355008 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.368981 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.380654 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ac1115e-d787-4ff7-b7f9-399137d327bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ca7d0113725fb7adcdea09bd4592580b9652ce2bb40e095484ec0c458203eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://069da5c66373778aec270d3597c665d396a164a6c07f258553405d7b269f3f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87d5fb88089c845017e6226c86fd1a41fc3622ad32b9ffec123ebb3649be4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.393472 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.404290 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.415927 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.429245 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.436609 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.436630 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.436639 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.436651 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.436660 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.439210 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.453324 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.470902 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.486512 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.503953 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.526015 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.539577 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.539627 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.539642 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.539664 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.539677 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.550505 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.567262 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.582708 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.599977 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:09Z\\\",\\\"message\\\":\\\"2025-11-28T11:22:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956\\\\n2025-11-28T11:22:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956 to /host/opt/cni/bin/\\\\n2025-11-28T11:22:24Z [verbose] multus-daemon started\\\\n2025-11-28T11:22:24Z [verbose] Readiness Indicator file check\\\\n2025-11-28T11:23:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.618289 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.633563 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:14Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.643290 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.643343 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.643362 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.643389 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.643406 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.746488 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.746533 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.746542 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.746557 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.746570 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.838443 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:14 crc kubenswrapper[4862]: E1128 11:23:14.838672 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.849373 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.849444 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.849461 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.849486 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.849531 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.952638 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.952694 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.952705 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.952722 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:14 crc kubenswrapper[4862]: I1128 11:23:14.952736 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:14Z","lastTransitionTime":"2025-11-28T11:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.056238 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.056322 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.056346 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.056371 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.056390 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.159186 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.159245 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.159263 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.159292 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.159313 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.263321 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.263394 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.263412 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.263441 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.263461 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.361340 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/3.log" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.362526 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/2.log" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.365651 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.365727 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.365744 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.365767 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.365784 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.367707 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" exitCode=1 Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.367761 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.367814 4862 scope.go:117] "RemoveContainer" containerID="e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.368967 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:23:15 crc kubenswrapper[4862]: E1128 11:23:15.369292 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.387580 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.404952 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.435703 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.462058 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.468213 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.468252 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.468270 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.468292 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.468309 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.476518 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.491260 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:09Z\\\",\\\"message\\\":\\\"2025-11-28T11:22:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956\\\\n2025-11-28T11:22:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956 to /host/opt/cni/bin/\\\\n2025-11-28T11:22:24Z [verbose] multus-daemon started\\\\n2025-11-28T11:22:24Z [verbose] Readiness Indicator file check\\\\n2025-11-28T11:23:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.509318 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.523380 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.538131 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.553627 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.566798 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ac1115e-d787-4ff7-b7f9-399137d327bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ca7d0113725fb7adcdea09bd4592580b9652ce2bb40e095484ec0c458203eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://069da5c66373778aec270d3597c665d396a164a6c07f258553405d7b269f3f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87d5fb88089c845017e6226c86fd1a41fc3622ad32b9ffec123ebb3649be4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.570668 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.570700 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.570708 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.570722 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.570733 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.585744 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.600115 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.618476 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.638609 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.654367 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.673117 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.673180 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.673200 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.673227 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.673244 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.673957 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.703431 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a4750c12da84a7c3b234b81f312c41cc16ba2b264bc9ca1c29a01371b7153f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:22:48Z\\\",\\\"message\\\":\\\"g reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.641960 6481 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642229 6481 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.642723 6481 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642839 6481 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642924 6481 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 11:22:48.642984 6481 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 11:22:48.643169 6481 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 11:22:48.643218 6481 factory.go:656] Stopping watch factory\\\\nI1128 11:22:48.643238 6481 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:14Z\\\",\\\"message\\\":\\\"834 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1128 11:23:14.741437 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1128 11:23:14.741454 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1128 11:23:14.741461 6834 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1128 11:23:14.741369 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 11:23:14.741483 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1128 11:23:14.741503 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:15Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.776055 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.776143 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.776161 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.776185 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.776202 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.838195 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.838283 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:15 crc kubenswrapper[4862]: E1128 11:23:15.838339 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.838290 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:15 crc kubenswrapper[4862]: E1128 11:23:15.838467 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:15 crc kubenswrapper[4862]: E1128 11:23:15.838628 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.878918 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.878969 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.878981 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.878998 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.879009 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.981304 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.981359 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.981376 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.981397 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:15 crc kubenswrapper[4862]: I1128 11:23:15.981412 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:15Z","lastTransitionTime":"2025-11-28T11:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.084886 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.084961 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.084984 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.085011 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.085030 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.188297 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.188352 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.188370 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.188396 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.188414 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.291433 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.291496 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.291513 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.291536 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.291552 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.374685 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/3.log" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.381447 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:23:16 crc kubenswrapper[4862]: E1128 11:23:16.381774 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.408043 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.408152 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.408244 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.408273 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.408291 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.413879 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:14Z\\\",\\\"message\\\":\\\"834 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1128 11:23:14.741437 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1128 11:23:14.741454 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1128 11:23:14.741461 6834 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1128 11:23:14.741369 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 11:23:14.741483 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1128 11:23:14.741503 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:23:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.433386 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.448954 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.470871 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.491703 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:09Z\\\",\\\"message\\\":\\\"2025-11-28T11:22:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956\\\\n2025-11-28T11:22:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956 to /host/opt/cni/bin/\\\\n2025-11-28T11:22:24Z [verbose] multus-daemon started\\\\n2025-11-28T11:22:24Z [verbose] Readiness Indicator file check\\\\n2025-11-28T11:23:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.505323 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.511333 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.511398 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.511417 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.511442 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.511459 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.519448 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.538393 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.538465 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.538491 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.538520 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.538537 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.553649 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: E1128 11:23:16.556589 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.562687 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.562883 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.563043 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.563244 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.563409 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.577792 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: E1128 11:23:16.584774 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.589821 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.589850 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.589861 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.589879 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.589892 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.602132 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: E1128 11:23:16.609617 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.613910 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.613945 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.613955 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.613973 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.613985 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.628524 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: E1128 11:23:16.636706 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.641905 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.642160 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.642367 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.642509 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.642640 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.658911 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: E1128 11:23:16.663392 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: E1128 11:23:16.663998 4862 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.666432 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.666476 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.666490 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.666508 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.666522 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.674130 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.691767 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.706791 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.724124 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ac1115e-d787-4ff7-b7f9-399137d327bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ca7d0113725fb7adcdea09bd4592580b9652ce2bb40e095484ec0c458203eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://069da5c66373778aec270d3597c665d396a164a6c07f258553405d7b269f3f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87d5fb88089c845017e6226c86fd1a41fc3622ad32b9ffec123ebb3649be4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.744350 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.765074 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:16Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.769891 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.769954 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.769972 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.770483 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.770556 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.838201 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:16 crc kubenswrapper[4862]: E1128 11:23:16.838394 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.874228 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.874274 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.874291 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.874314 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.874332 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.976955 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.976994 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.977009 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.977030 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:16 crc kubenswrapper[4862]: I1128 11:23:16.977046 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:16Z","lastTransitionTime":"2025-11-28T11:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.080027 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.080070 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.080086 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.080138 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.080154 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:17Z","lastTransitionTime":"2025-11-28T11:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.182534 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.182582 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.182598 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.182619 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.182635 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:17Z","lastTransitionTime":"2025-11-28T11:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.285660 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.285712 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.285733 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.285760 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.285784 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:17Z","lastTransitionTime":"2025-11-28T11:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.388550 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.388606 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.388625 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.388646 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.388662 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:17Z","lastTransitionTime":"2025-11-28T11:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.490937 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.491314 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.491455 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.491604 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.491813 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:17Z","lastTransitionTime":"2025-11-28T11:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.595719 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.595767 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.595783 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.595805 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.595821 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:17Z","lastTransitionTime":"2025-11-28T11:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.699148 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.699607 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.699789 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.700034 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.700300 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:17Z","lastTransitionTime":"2025-11-28T11:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.804214 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.804294 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.804320 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.804376 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.804403 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:17Z","lastTransitionTime":"2025-11-28T11:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.837953 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.837983 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.837964 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:17 crc kubenswrapper[4862]: E1128 11:23:17.838134 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:17 crc kubenswrapper[4862]: E1128 11:23:17.838302 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:17 crc kubenswrapper[4862]: E1128 11:23:17.838421 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.856874 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.907967 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.908037 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.908059 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.908087 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:17 crc kubenswrapper[4862]: I1128 11:23:17.908159 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:17Z","lastTransitionTime":"2025-11-28T11:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.011335 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.011772 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.011918 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.012083 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.012298 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.115949 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.116002 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.116019 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.116044 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.116063 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.219912 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.219972 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.219996 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.220044 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.220065 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.323787 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.323843 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.323866 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.323899 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.323924 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.427226 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.427282 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.427299 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.427325 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.427362 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.530662 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.530722 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.530747 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.530774 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.530797 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.633521 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.633833 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.633924 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.634034 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.634137 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.736699 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.736854 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.736936 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.737029 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.737134 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.837871 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:18 crc kubenswrapper[4862]: E1128 11:23:18.838172 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.839317 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.839372 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.839392 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.839415 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.839432 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.942594 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.942653 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.942678 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.942705 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:18 crc kubenswrapper[4862]: I1128 11:23:18.942727 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:18Z","lastTransitionTime":"2025-11-28T11:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.046240 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.046303 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.046329 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.046358 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.046381 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.149609 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.149855 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.150144 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.150379 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.150619 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.254196 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.254260 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.254281 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.254309 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.254331 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.357141 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.357208 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.357225 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.357254 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.357273 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.460568 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.460625 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.460647 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.460671 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.460688 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.564001 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.564048 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.564067 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.564120 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.564145 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.667633 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.667699 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.667747 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.667773 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.667789 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.770351 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.770461 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.770491 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.770527 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.770551 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.838344 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.838390 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:19 crc kubenswrapper[4862]: E1128 11:23:19.838825 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.838463 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:19 crc kubenswrapper[4862]: E1128 11:23:19.839152 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:19 crc kubenswrapper[4862]: E1128 11:23:19.839442 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.874181 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.874246 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.874269 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.874298 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.874320 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.977551 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.977618 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.977636 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.977659 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:19 crc kubenswrapper[4862]: I1128 11:23:19.977675 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:19Z","lastTransitionTime":"2025-11-28T11:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.080543 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.080601 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.080619 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.080646 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.080668 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:20Z","lastTransitionTime":"2025-11-28T11:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.184800 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.184860 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.184877 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.184903 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.184926 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:20Z","lastTransitionTime":"2025-11-28T11:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.288237 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.288301 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.288319 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.288345 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.288363 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:20Z","lastTransitionTime":"2025-11-28T11:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.392182 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.392282 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.392307 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.392339 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.392362 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:20Z","lastTransitionTime":"2025-11-28T11:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.496171 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.496232 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.496249 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.496278 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.496297 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:20Z","lastTransitionTime":"2025-11-28T11:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.599492 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.599562 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.599580 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.599608 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.599626 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:20Z","lastTransitionTime":"2025-11-28T11:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.702817 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.702877 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.702894 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.702915 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.702932 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:20Z","lastTransitionTime":"2025-11-28T11:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.806577 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.806957 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.807174 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.807338 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.807468 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:20Z","lastTransitionTime":"2025-11-28T11:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.838234 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:20 crc kubenswrapper[4862]: E1128 11:23:20.838425 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.910042 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.910198 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.910225 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.910255 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:20 crc kubenswrapper[4862]: I1128 11:23:20.910279 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:20Z","lastTransitionTime":"2025-11-28T11:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.014282 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.014349 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.014367 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.014393 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.014411 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.118322 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.118396 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.118423 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.118454 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.118477 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.222076 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.222254 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.222275 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.222299 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.222316 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.326525 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.326622 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.326640 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.326694 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.326712 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.431065 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.431224 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.431287 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.431318 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.431376 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.535725 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.536240 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.536403 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.536560 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.536703 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.640456 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.640578 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.640599 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.640668 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.640731 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.744629 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.744716 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.744735 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.744765 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.744786 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.838443 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.838478 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.838611 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:21 crc kubenswrapper[4862]: E1128 11:23:21.838802 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:21 crc kubenswrapper[4862]: E1128 11:23:21.839035 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:21 crc kubenswrapper[4862]: E1128 11:23:21.839183 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.848623 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.848686 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.848706 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.848734 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.848754 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.952798 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.952875 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.952901 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.952931 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:21 crc kubenswrapper[4862]: I1128 11:23:21.952959 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:21Z","lastTransitionTime":"2025-11-28T11:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.057278 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.057339 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.057358 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.057385 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.057403 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.160721 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.160782 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.160798 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.160821 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.160839 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.264514 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.264953 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.265373 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.265621 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.265840 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.370068 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.370776 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.370856 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.370929 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.370998 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.473639 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.473724 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.473738 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.473755 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.473770 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.577670 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.577747 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.577765 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.577799 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.577823 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.682143 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.682211 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.682228 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.682256 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.682302 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.785030 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.785136 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.785164 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.785196 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.785218 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.838385 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:22 crc kubenswrapper[4862]: E1128 11:23:22.838545 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.854004 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3aa0652-2a33-4a6f-88b3-4576b5dd6794\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f59dd4bb9483584434a420817fc9b55580712d43b6e8267e286e1b6e3acb83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01fb725ec46ee8b2588256779d17793f286445d63eb0e0a2acef74beac25678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01fb725ec46ee8b2588256779d17793f286445d63eb0e0a2acef74beac25678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.874634 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ac1115e-d787-4ff7-b7f9-399137d327bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ca7d0113725fb7adcdea09bd4592580b9652ce2bb40e095484ec0c458203eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://069da5c66373778aec270d3597c665d396a164a6c07f258553405d7b269f3f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87d5fb88089c845017e6226c86fd1a41fc3622ad32b9ffec123ebb3649be4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1d50f17371336578073c838bc30c6f104f7137b05f0a6e313316d584634bfb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.891314 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.891363 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.891379 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.891405 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.891422 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.896357 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd84dc2-86ad-4cc8-b648-1a995db185db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a5916682a4f1372d969637297c112e1155b511dbdb4f1840ea7a32fe7882087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://845774c8903d5c82deead5e0a7cf4a5c001f19ae4daa616538d2475776b9627e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4d981366bbd5f06164ad21e8d14ae5f4016e09b93e3408be4fbbbb83f74c17d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.917137 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3561a641744436d04ac7e39a42e178701f9cc255ccc3d7f1787135ddc14fad7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e125b864e22af3e8f0514d1b2f961c11664adc56bceb944e120e9a11bcf4a660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.940385 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.965529 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406a4ee3bff5ffe5a44ee96c101cdd250001e9f15c875acc28a1a11380767d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.981565 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-25clg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6de873-3ea5-4223-a260-9451e5e6e803\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefad6a343b767ad7669134f8e778b6b7d7509d48538cafdadcf738de54c6d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlwp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-25clg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.995373 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.995441 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.995466 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.995495 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.995518 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:22Z","lastTransitionTime":"2025-11-28T11:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:22 crc kubenswrapper[4862]: I1128 11:23:22.997204 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6845439-1a4c-4f28-9ba4-29149d545995\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87ef50302eb2f85535e814b3cb62e8b88b3fc0d938187cabe6422502e14c38d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2p4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fm8j2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:22Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.011797 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60f0df65-b253-4ea2-abda-dbc50f7f2eca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7tgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl4cj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.079237 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.096115 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87f7efa62365b9b2fbba19b33a2c45d55192c7977acb599d26cdb44e37a320f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.098410 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.098481 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.098504 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.098535 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.098557 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:23Z","lastTransitionTime":"2025-11-28T11:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.122705 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68a1bd24-7d7d-4d41-9f82-ff6158c145d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62d8ec87cc6de8fdd4b7844503c5206656cdc4d10e622021e36b5a6f4ab58f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb84f077f19ad9046fbde10f7424a4f815a7cca461f89b26c1dda79793954276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6932c9b529b6e9e80738c60116b684e1c875509b14fd4cd66e0d36a78cb5af0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7340a5497385e79c6283e141fd9641227df0ad3beecf855563d3e4494ef1ff09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50fb37a728eccf22e03d2096651989ecaaa42cb4d7ac1b9dabf3af1ed1b99e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba0a8e30b7036ed909158589b60fda50c4892bf8ad3e09ffb559c73d679035cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33cf0d01005c73ba25c8ec4356f94946f41d268e8fc4dd449ec52028e64be45c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sq7pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2nh6f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.154961 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9916d783-c2df-47fa-9c38-ee7537db24d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:14Z\\\",\\\"message\\\":\\\"834 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1128 11:23:14.741437 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1128 11:23:14.741454 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1128 11:23:14.741461 6834 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1128 11:23:14.741369 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 11:23:14.741483 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1128 11:23:14.741503 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:23:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjwcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ccpp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.189735 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8b7931-9129-472b-9472-2e5aa2e92938\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fad7fead8f086ce76d22a58cdf2d6682c57c3dd41dbb13bf4ecdd033305a1b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d87915e25e5c7dfb67d2afbd14af78770ae1b67375dd79e02bec2bbb02529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c6488a029bdc206243a0365b8460a6206170ac623e8e18efaa8e740ba9a6102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c87434ef7db0aa8fcd837be953cf8af654413837144dfc32977eb0f8dbb62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb6a958883bf69e1c6cdbc91a77685ba6a7242b5319a702ca97d81b3c2d7200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1b8d393aac9a870cd886084bc53453e87c8d95199595cd272e556e296cb93e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82a5c2013a717b9b21472424237f5acef92dd954331d53645da419d5177b17b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dac93cc7f5b746e505aef035ff605efd8c0af8c51211edab9306e4e214d0732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.201742 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.202148 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.202309 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.202544 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.202712 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:23Z","lastTransitionTime":"2025-11-28T11:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.214571 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0660f02d-24c4-419c-a304-0401773eed51\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 11:22:15.383956 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 11:22:15.385944 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3882954516/tls.crt::/tmp/serving-cert-3882954516/tls.key\\\\\\\"\\\\nI1128 11:22:21.285415 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 11:22:21.289720 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 11:22:21.289758 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 11:22:21.289822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 11:22:21.289842 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 11:22:21.298488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 11:22:21.298521 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298531 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 11:22:21.298540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 11:22:21.298546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 11:22:21.298552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 11:22:21.298579 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 11:22:21.298818 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 11:22:21.303969 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T11:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.237374 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.258633 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kqpd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6abffd5-6086-4a32-ba8d-e7c669548c5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T11:23:09Z\\\",\\\"message\\\":\\\"2025-11-28T11:22:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956\\\\n2025-11-28T11:22:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2c7f989-e4bd-4179-b792-33346bdb3956 to /host/opt/cni/bin/\\\\n2025-11-28T11:22:24Z [verbose] multus-daemon started\\\\n2025-11-28T11:22:24Z [verbose] Readiness Indicator file check\\\\n2025-11-28T11:23:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T11:22:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59vmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kqpd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.274975 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tpk22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f84710d-eb96-4879-8704-1e8dda17d4b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d893dad69623a1bb4801f3e4c6b7d807df03f4371c8492d4c912d6501fc062c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b99r6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tpk22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.291950 4862 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dfd45e3-0af1-469f-ab6a-4fd5ff606718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T11:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7c05b18685dc7080e6c93008349d5350c0e5240114f87c6b84d225f61a92386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa81ba1bbab8b267f16a5642f1e747a05d079219d1b4c7ff766bc494198d3ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T11:22:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwcqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T11:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lg2cn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:23Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.305715 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.305950 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.306168 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.306414 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.306619 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:23Z","lastTransitionTime":"2025-11-28T11:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.408816 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.408883 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.408902 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.408929 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.408950 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:23Z","lastTransitionTime":"2025-11-28T11:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.511679 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.511720 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.511732 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.511748 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.511760 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:23Z","lastTransitionTime":"2025-11-28T11:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.615573 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.615647 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.615669 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.615697 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.615714 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:23Z","lastTransitionTime":"2025-11-28T11:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.719465 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.719539 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.719551 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.719575 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.719591 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:23Z","lastTransitionTime":"2025-11-28T11:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.823219 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.823293 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.823307 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.823334 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.823355 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:23Z","lastTransitionTime":"2025-11-28T11:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.838504 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.838632 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.838529 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:23 crc kubenswrapper[4862]: E1128 11:23:23.838713 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:23 crc kubenswrapper[4862]: E1128 11:23:23.838867 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:23 crc kubenswrapper[4862]: E1128 11:23:23.839401 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.930235 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.930290 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.930309 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.930341 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:23 crc kubenswrapper[4862]: I1128 11:23:23.930365 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:23Z","lastTransitionTime":"2025-11-28T11:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.033757 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.033822 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.033840 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.033863 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.033880 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.136967 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.137032 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.137050 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.137074 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.137119 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.244717 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.244827 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.244849 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.244885 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.244917 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.349319 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.349379 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.349396 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.349421 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.349439 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.452940 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.453043 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.453140 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.453175 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.453243 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.556744 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.556823 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.556847 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.556875 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.556894 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.660764 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.660824 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.660841 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.660864 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.660881 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.764248 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.764312 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.764330 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.764358 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.764376 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.838686 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:24 crc kubenswrapper[4862]: E1128 11:23:24.838934 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.867942 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.868226 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.868402 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.868556 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.868743 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.972566 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.972629 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.972648 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.972673 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:24 crc kubenswrapper[4862]: I1128 11:23:24.972691 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:24Z","lastTransitionTime":"2025-11-28T11:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.075393 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.075454 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.075477 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.075505 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.075527 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:25Z","lastTransitionTime":"2025-11-28T11:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.178322 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.178371 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.178387 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.178408 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.178428 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:25Z","lastTransitionTime":"2025-11-28T11:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.281551 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.281620 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.281646 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.281677 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.281696 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:25Z","lastTransitionTime":"2025-11-28T11:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.384580 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.384647 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.384665 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.384689 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.384713 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:25Z","lastTransitionTime":"2025-11-28T11:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.487510 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.487573 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.487596 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.487624 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.487646 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:25Z","lastTransitionTime":"2025-11-28T11:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.590836 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.590898 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.590917 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.590941 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.590960 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:25Z","lastTransitionTime":"2025-11-28T11:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.693317 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.693727 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.693942 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.694178 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.694389 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:25Z","lastTransitionTime":"2025-11-28T11:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.796308 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.796671 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.796645144 +0000 UTC m=+147.329159105 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.797935 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.797994 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.798019 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.798045 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.798067 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:25Z","lastTransitionTime":"2025-11-28T11:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.838072 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.838127 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.838300 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.838569 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.838707 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.839359 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.897451 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.897502 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.897551 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.897589 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897690 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897725 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897729 4862 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897775 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897797 4862 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897813 4862 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897848 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.897817215 +0000 UTC m=+147.430331176 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897745 4862 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897879 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.897865857 +0000 UTC m=+147.430379808 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897700 4862 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897904 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.897893197 +0000 UTC m=+147.430407148 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 11:23:25 crc kubenswrapper[4862]: E1128 11:23:25.897935 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.897915338 +0000 UTC m=+147.430429299 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.900713 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.900759 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.900776 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.900798 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:25 crc kubenswrapper[4862]: I1128 11:23:25.900813 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:25Z","lastTransitionTime":"2025-11-28T11:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.004544 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.004610 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.004632 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.004655 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.004671 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.108460 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.108513 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.108534 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.108560 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.108581 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.211717 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.211798 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.211822 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.211850 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.211867 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.314656 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.314726 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.314744 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.314772 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.314792 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.419244 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.419305 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.419327 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.419355 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.419378 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.522578 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.522670 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.522682 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.522706 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.522717 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.625647 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.625701 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.625719 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.625739 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.625753 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.728389 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.728448 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.728469 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.728496 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.728514 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.834018 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.834335 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.834362 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.834395 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.834418 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.838571 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:26 crc kubenswrapper[4862]: E1128 11:23:26.838772 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.869753 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.869811 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.869824 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.869843 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.869856 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: E1128 11:23:26.890272 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.896182 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.896243 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.896258 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.896277 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.896291 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: E1128 11:23:26.910556 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.917175 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.917213 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.917228 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.917248 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.917263 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: E1128 11:23:26.938772 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.944853 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.944899 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.944912 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.944937 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.944952 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: E1128 11:23:26.961920 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.967201 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.967295 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.967312 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.967334 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.967355 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:26 crc kubenswrapper[4862]: E1128 11:23:26.985504 4862 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T11:23:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c620f82-404c-4d65-9ba0-65e90c64ab19\\\",\\\"systemUUID\\\":\\\"b1809799-f6d1-48c5-b4d0-d4f1bd5f5602\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T11:23:26Z is after 2025-08-24T17:21:41Z" Nov 28 11:23:26 crc kubenswrapper[4862]: E1128 11:23:26.985689 4862 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.988037 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.988113 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.988133 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.988164 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:26 crc kubenswrapper[4862]: I1128 11:23:26.988183 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:26Z","lastTransitionTime":"2025-11-28T11:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.091262 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.091311 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.091324 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.091344 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.091358 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:27Z","lastTransitionTime":"2025-11-28T11:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.194421 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.194483 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.194492 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.194512 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.194525 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:27Z","lastTransitionTime":"2025-11-28T11:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.297297 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.297365 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.297386 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.297412 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.297432 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:27Z","lastTransitionTime":"2025-11-28T11:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.401454 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.401956 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.401974 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.402001 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.402019 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:27Z","lastTransitionTime":"2025-11-28T11:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.505736 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.505791 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.505808 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.505830 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.505847 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:27Z","lastTransitionTime":"2025-11-28T11:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.610523 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.610811 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.610833 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.610857 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.610877 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:27Z","lastTransitionTime":"2025-11-28T11:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.713711 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.713778 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.713796 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.713820 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.713838 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:27Z","lastTransitionTime":"2025-11-28T11:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.817624 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.817685 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.817702 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.817726 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.817746 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:27Z","lastTransitionTime":"2025-11-28T11:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.837997 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.838133 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.838003 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:27 crc kubenswrapper[4862]: E1128 11:23:27.838222 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:27 crc kubenswrapper[4862]: E1128 11:23:27.838481 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:27 crc kubenswrapper[4862]: E1128 11:23:27.838585 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.921753 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.921850 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.921884 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.921913 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:27 crc kubenswrapper[4862]: I1128 11:23:27.921943 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:27Z","lastTransitionTime":"2025-11-28T11:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.025272 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.025323 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.025339 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.025364 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.025380 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.128200 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.128261 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.128281 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.128305 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.128324 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.230852 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.230918 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.230937 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.230962 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.230979 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.334849 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.334913 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.334934 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.334963 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.334987 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.437179 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.437589 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.437698 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.437907 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.438014 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.540947 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.541491 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.541674 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.541847 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.542027 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.645581 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.645664 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.645686 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.645710 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.645727 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.748758 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.748859 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.748881 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.748911 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.748934 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.838044 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:28 crc kubenswrapper[4862]: E1128 11:23:28.838321 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.840195 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:23:28 crc kubenswrapper[4862]: E1128 11:23:28.840757 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.851916 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.851998 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.852025 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.852054 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.852079 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.955717 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.955883 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.955907 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.955933 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:28 crc kubenswrapper[4862]: I1128 11:23:28.955951 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:28Z","lastTransitionTime":"2025-11-28T11:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.058752 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.058837 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.058860 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.058883 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.058901 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.161585 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.161633 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.161651 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.161673 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.161690 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.265387 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.265456 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.265474 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.265500 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.265517 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.368655 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.368716 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.368749 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.368803 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.368832 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.471963 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.472022 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.472043 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.472066 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.472084 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.574430 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.574495 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.574535 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.574655 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.574680 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.677337 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.677390 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.677406 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.677428 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.677446 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.780415 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.780572 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.780596 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.780620 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.780637 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.838060 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.838124 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:29 crc kubenswrapper[4862]: E1128 11:23:29.838361 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.838139 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:29 crc kubenswrapper[4862]: E1128 11:23:29.838482 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:29 crc kubenswrapper[4862]: E1128 11:23:29.838696 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.884886 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.884959 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.884979 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.885007 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.885025 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.988214 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.988279 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.988296 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.988319 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:29 crc kubenswrapper[4862]: I1128 11:23:29.988336 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:29Z","lastTransitionTime":"2025-11-28T11:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.091233 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.091302 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.091325 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.091353 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.091374 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:30Z","lastTransitionTime":"2025-11-28T11:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.194982 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.195057 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.195080 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.195144 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.195167 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:30Z","lastTransitionTime":"2025-11-28T11:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.299260 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.299341 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.299364 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.299394 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.299416 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:30Z","lastTransitionTime":"2025-11-28T11:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.402525 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.402596 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.402613 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.402640 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.402659 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:30Z","lastTransitionTime":"2025-11-28T11:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.506221 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.506375 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.506388 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.506413 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.506428 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:30Z","lastTransitionTime":"2025-11-28T11:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.610379 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.610555 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.610580 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.610664 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.611034 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:30Z","lastTransitionTime":"2025-11-28T11:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.715596 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.715691 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.715707 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.716218 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.716255 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:30Z","lastTransitionTime":"2025-11-28T11:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.820309 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.820364 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.820395 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.820419 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.820436 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:30Z","lastTransitionTime":"2025-11-28T11:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.838207 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:30 crc kubenswrapper[4862]: E1128 11:23:30.838381 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.923638 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.923720 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.923734 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.923756 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:30 crc kubenswrapper[4862]: I1128 11:23:30.923770 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:30Z","lastTransitionTime":"2025-11-28T11:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.027061 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.027164 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.027188 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.027218 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.027241 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.130954 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.131143 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.131163 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.131192 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.131209 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.233873 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.233964 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.233987 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.234019 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.234045 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.337215 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.337259 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.337279 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.337301 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.337318 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.440182 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.440265 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.440288 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.440322 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.440344 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.542960 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.543134 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.543155 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.543182 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.543204 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.645746 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.645811 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.645819 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.645835 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.645843 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.748802 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.748846 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.748856 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.748871 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.748880 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.838526 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.838548 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:31 crc kubenswrapper[4862]: E1128 11:23:31.838773 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:31 crc kubenswrapper[4862]: E1128 11:23:31.838940 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.838593 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:31 crc kubenswrapper[4862]: E1128 11:23:31.839334 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.852986 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.853050 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.853069 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.853135 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.853155 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.955395 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.955433 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.955442 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.955456 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:31 crc kubenswrapper[4862]: I1128 11:23:31.955464 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:31Z","lastTransitionTime":"2025-11-28T11:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.058562 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.058631 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.058649 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.058674 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.058695 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.161428 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.161487 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.161510 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.161535 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.161554 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.264170 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.264267 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.264284 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.264344 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.264363 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.367854 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.367914 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.367931 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.367953 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.367970 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.470446 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.470499 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.470515 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.470537 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.470554 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.572842 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.572911 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.572933 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.572960 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.572980 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.676221 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.676281 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.676299 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.676323 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.676345 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.779824 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.779878 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.779898 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.779925 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.779944 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.838020 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:32 crc kubenswrapper[4862]: E1128 11:23:32.838201 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.884463 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.884528 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.884550 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.884584 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.884607 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.974490 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2nh6f" podStartSLOduration=70.974459279 podStartE2EDuration="1m10.974459279s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:32.921137525 +0000 UTC m=+90.453651476" watchObservedRunningTime="2025-11-28 11:23:32.974459279 +0000 UTC m=+90.506973240" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.987930 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.987991 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.988008 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.988033 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:32 crc kubenswrapper[4862]: I1128 11:23:32.988050 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:32Z","lastTransitionTime":"2025-11-28T11:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.013545 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=70.013526727 podStartE2EDuration="1m10.013526727s" podCreationTimestamp="2025-11-28 11:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.012109073 +0000 UTC m=+90.544623004" watchObservedRunningTime="2025-11-28 11:23:33.013526727 +0000 UTC m=+90.546040668" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.039359 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=72.039324168 podStartE2EDuration="1m12.039324168s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.037814491 +0000 UTC m=+90.570328422" watchObservedRunningTime="2025-11-28 11:23:33.039324168 +0000 UTC m=+90.571838139" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.071197 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-kqpd4" podStartSLOduration=71.071177344 podStartE2EDuration="1m11.071177344s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.069978847 +0000 UTC m=+90.602492778" watchObservedRunningTime="2025-11-28 11:23:33.071177344 +0000 UTC m=+90.603691275" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.087808 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tpk22" podStartSLOduration=72.087788323 podStartE2EDuration="1m12.087788323s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.086925577 +0000 UTC m=+90.619439538" watchObservedRunningTime="2025-11-28 11:23:33.087788323 +0000 UTC m=+90.620302254" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.090756 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.090959 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.091084 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.091285 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.091398 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:33Z","lastTransitionTime":"2025-11-28T11:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.126067 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lg2cn" podStartSLOduration=71.126041126 podStartE2EDuration="1m11.126041126s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.124707945 +0000 UTC m=+90.657221886" watchObservedRunningTime="2025-11-28 11:23:33.126041126 +0000 UTC m=+90.658555087" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.138974 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=16.138956552 podStartE2EDuration="16.138956552s" podCreationTimestamp="2025-11-28 11:23:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.1379306 +0000 UTC m=+90.670444541" watchObservedRunningTime="2025-11-28 11:23:33.138956552 +0000 UTC m=+90.671470483" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.155790 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podStartSLOduration=71.155758937 podStartE2EDuration="1m11.155758937s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.154501068 +0000 UTC m=+90.687015029" watchObservedRunningTime="2025-11-28 11:23:33.155758937 +0000 UTC m=+90.688272898" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.193995 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.194060 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.194077 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.194119 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.194136 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:33Z","lastTransitionTime":"2025-11-28T11:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.210795 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=37.210772813 podStartE2EDuration="37.210772813s" podCreationTimestamp="2025-11-28 11:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.194277607 +0000 UTC m=+90.726791528" watchObservedRunningTime="2025-11-28 11:23:33.210772813 +0000 UTC m=+90.743286774" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.239324 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=72.239301967 podStartE2EDuration="1m12.239301967s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.215754236 +0000 UTC m=+90.748268157" watchObservedRunningTime="2025-11-28 11:23:33.239301967 +0000 UTC m=+90.771815918" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.285857 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-25clg" podStartSLOduration=72.285841864 podStartE2EDuration="1m12.285841864s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:33.285195894 +0000 UTC m=+90.817709815" watchObservedRunningTime="2025-11-28 11:23:33.285841864 +0000 UTC m=+90.818355785" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.296944 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.297009 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.297026 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.297051 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.297068 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:33Z","lastTransitionTime":"2025-11-28T11:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.400534 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.400589 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.400611 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.400646 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.400665 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:33Z","lastTransitionTime":"2025-11-28T11:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.503905 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.503973 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.503992 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.504018 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.504039 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:33Z","lastTransitionTime":"2025-11-28T11:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.606956 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.607021 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.607044 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.607073 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.607144 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:33Z","lastTransitionTime":"2025-11-28T11:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.710492 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.710562 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.710580 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.710605 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.710626 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:33Z","lastTransitionTime":"2025-11-28T11:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.813611 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.813684 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.813708 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.813737 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.813762 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:33Z","lastTransitionTime":"2025-11-28T11:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.838569 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.838765 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:33 crc kubenswrapper[4862]: E1128 11:23:33.839023 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.839055 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:33 crc kubenswrapper[4862]: E1128 11:23:33.839297 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:33 crc kubenswrapper[4862]: E1128 11:23:33.839443 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.916692 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.916756 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.916773 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.916808 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:33 crc kubenswrapper[4862]: I1128 11:23:33.916827 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:33Z","lastTransitionTime":"2025-11-28T11:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.020420 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.020516 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.020548 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.020579 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.020600 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.123030 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.123086 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.123129 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.123152 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.123171 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.225594 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.225659 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.225677 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.225699 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.225715 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.335388 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.335443 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.335473 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.335493 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.335507 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.438241 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.438332 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.438352 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.438379 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.438398 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.540937 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.540995 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.541011 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.541034 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.541054 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.643722 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.643782 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.643802 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.643831 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.643848 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.745969 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.746032 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.746048 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.746076 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.746121 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.837955 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:34 crc kubenswrapper[4862]: E1128 11:23:34.838188 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.848387 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.848484 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.848503 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.848537 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.848559 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.951629 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.951687 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.951705 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.951729 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:34 crc kubenswrapper[4862]: I1128 11:23:34.951746 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:34Z","lastTransitionTime":"2025-11-28T11:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.055421 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.055511 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.055541 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.055576 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.055602 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.158443 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.158519 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.158537 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.158563 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.158581 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.262011 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.262083 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.262143 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.262173 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.262193 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.365204 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.365286 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.365307 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.365332 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.365352 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.468227 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.468297 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.468319 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.468368 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.468397 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.571560 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.571635 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.571659 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.571688 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.571709 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.675601 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.675683 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.675705 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.675740 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.675763 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.780190 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.780260 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.780282 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.780316 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.780338 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.837956 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:35 crc kubenswrapper[4862]: E1128 11:23:35.838164 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.838376 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:35 crc kubenswrapper[4862]: E1128 11:23:35.838549 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.838620 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:35 crc kubenswrapper[4862]: E1128 11:23:35.838774 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.883567 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.883650 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.883675 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.883704 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.883729 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.986611 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.986693 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.986714 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.986750 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:35 crc kubenswrapper[4862]: I1128 11:23:35.986771 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:35Z","lastTransitionTime":"2025-11-28T11:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.089908 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.089976 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.089993 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.090017 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.090035 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:36Z","lastTransitionTime":"2025-11-28T11:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.192881 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.192942 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.192958 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.192980 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.193003 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:36Z","lastTransitionTime":"2025-11-28T11:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.296180 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.296237 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.296287 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.296311 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.296328 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:36Z","lastTransitionTime":"2025-11-28T11:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.400207 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.400265 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.400283 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.400307 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.400324 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:36Z","lastTransitionTime":"2025-11-28T11:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.504525 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.504591 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.504610 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.504634 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.504653 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:36Z","lastTransitionTime":"2025-11-28T11:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.607511 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.607573 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.607592 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.607617 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.607638 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:36Z","lastTransitionTime":"2025-11-28T11:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.712420 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.712480 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.712500 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.712525 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.712542 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:36Z","lastTransitionTime":"2025-11-28T11:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.816655 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.816743 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.816784 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.816817 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.816841 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:36Z","lastTransitionTime":"2025-11-28T11:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.837770 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:36 crc kubenswrapper[4862]: E1128 11:23:36.837994 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.919585 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.919649 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.919666 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.919691 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:36 crc kubenswrapper[4862]: I1128 11:23:36.919710 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:36Z","lastTransitionTime":"2025-11-28T11:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.023687 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.023740 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.023753 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.023778 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.023793 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:37Z","lastTransitionTime":"2025-11-28T11:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.127229 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.127652 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.128294 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.128355 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.128370 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:37Z","lastTransitionTime":"2025-11-28T11:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.231211 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.231267 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.231279 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.231308 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.231322 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:37Z","lastTransitionTime":"2025-11-28T11:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.286008 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.286066 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.286076 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.286131 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.286146 4862 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T11:23:37Z","lastTransitionTime":"2025-11-28T11:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.345068 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr"] Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.346065 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.348009 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.349915 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.350444 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.350804 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.434932 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/135456f3-b2f6-4423-8fb5-6bc24c2713cd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.435037 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/135456f3-b2f6-4423-8fb5-6bc24c2713cd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.435080 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/135456f3-b2f6-4423-8fb5-6bc24c2713cd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.435146 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/135456f3-b2f6-4423-8fb5-6bc24c2713cd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.435171 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/135456f3-b2f6-4423-8fb5-6bc24c2713cd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.536715 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/135456f3-b2f6-4423-8fb5-6bc24c2713cd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.536785 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/135456f3-b2f6-4423-8fb5-6bc24c2713cd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.536874 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/135456f3-b2f6-4423-8fb5-6bc24c2713cd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.537003 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/135456f3-b2f6-4423-8fb5-6bc24c2713cd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.537176 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/135456f3-b2f6-4423-8fb5-6bc24c2713cd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.537236 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/135456f3-b2f6-4423-8fb5-6bc24c2713cd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.537243 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/135456f3-b2f6-4423-8fb5-6bc24c2713cd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.538995 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/135456f3-b2f6-4423-8fb5-6bc24c2713cd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.552714 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/135456f3-b2f6-4423-8fb5-6bc24c2713cd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.570464 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/135456f3-b2f6-4423-8fb5-6bc24c2713cd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-65jgr\" (UID: \"135456f3-b2f6-4423-8fb5-6bc24c2713cd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.673351 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" Nov 28 11:23:37 crc kubenswrapper[4862]: W1128 11:23:37.699221 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod135456f3_b2f6_4423_8fb5_6bc24c2713cd.slice/crio-4ab8d9cd43f30832ed3bde2a5a9ee2f0b0c1145065a846e988217e6f8096d0e2 WatchSource:0}: Error finding container 4ab8d9cd43f30832ed3bde2a5a9ee2f0b0c1145065a846e988217e6f8096d0e2: Status 404 returned error can't find the container with id 4ab8d9cd43f30832ed3bde2a5a9ee2f0b0c1145065a846e988217e6f8096d0e2 Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.837628 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.837726 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:37 crc kubenswrapper[4862]: E1128 11:23:37.837893 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:37 crc kubenswrapper[4862]: I1128 11:23:37.837964 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:37 crc kubenswrapper[4862]: E1128 11:23:37.838052 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:37 crc kubenswrapper[4862]: E1128 11:23:37.838354 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:38 crc kubenswrapper[4862]: I1128 11:23:38.475660 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" event={"ID":"135456f3-b2f6-4423-8fb5-6bc24c2713cd","Type":"ContainerStarted","Data":"0f138e8c166a859e8d2d9bf11e28f68d5f96abe57a45f844d70607928e794834"} Nov 28 11:23:38 crc kubenswrapper[4862]: I1128 11:23:38.475734 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" event={"ID":"135456f3-b2f6-4423-8fb5-6bc24c2713cd","Type":"ContainerStarted","Data":"4ab8d9cd43f30832ed3bde2a5a9ee2f0b0c1145065a846e988217e6f8096d0e2"} Nov 28 11:23:38 crc kubenswrapper[4862]: I1128 11:23:38.498341 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-65jgr" podStartSLOduration=76.498318748 podStartE2EDuration="1m16.498318748s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:23:38.497269236 +0000 UTC m=+96.029783167" watchObservedRunningTime="2025-11-28 11:23:38.498318748 +0000 UTC m=+96.030832709" Nov 28 11:23:38 crc kubenswrapper[4862]: I1128 11:23:38.838207 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:38 crc kubenswrapper[4862]: E1128 11:23:38.838451 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:39 crc kubenswrapper[4862]: I1128 11:23:39.838166 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:39 crc kubenswrapper[4862]: I1128 11:23:39.838270 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:39 crc kubenswrapper[4862]: E1128 11:23:39.838315 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:39 crc kubenswrapper[4862]: E1128 11:23:39.838452 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:39 crc kubenswrapper[4862]: I1128 11:23:39.838549 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:39 crc kubenswrapper[4862]: E1128 11:23:39.838628 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:39 crc kubenswrapper[4862]: I1128 11:23:39.839577 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:23:39 crc kubenswrapper[4862]: E1128 11:23:39.839827 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" Nov 28 11:23:40 crc kubenswrapper[4862]: I1128 11:23:40.774544 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:40 crc kubenswrapper[4862]: E1128 11:23:40.774707 4862 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:23:40 crc kubenswrapper[4862]: E1128 11:23:40.774763 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs podName:60f0df65-b253-4ea2-abda-dbc50f7f2eca nodeName:}" failed. No retries permitted until 2025-11-28 11:24:44.774750425 +0000 UTC m=+162.307264346 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs") pod "network-metrics-daemon-wl4cj" (UID: "60f0df65-b253-4ea2-abda-dbc50f7f2eca") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 11:23:40 crc kubenswrapper[4862]: I1128 11:23:40.838079 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:40 crc kubenswrapper[4862]: E1128 11:23:40.838328 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:41 crc kubenswrapper[4862]: I1128 11:23:41.838563 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:41 crc kubenswrapper[4862]: E1128 11:23:41.838777 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:41 crc kubenswrapper[4862]: I1128 11:23:41.839220 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:41 crc kubenswrapper[4862]: E1128 11:23:41.839354 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:41 crc kubenswrapper[4862]: I1128 11:23:41.839460 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:41 crc kubenswrapper[4862]: E1128 11:23:41.839667 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:42 crc kubenswrapper[4862]: I1128 11:23:42.837987 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:42 crc kubenswrapper[4862]: E1128 11:23:42.839605 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:43 crc kubenswrapper[4862]: I1128 11:23:43.837871 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:43 crc kubenswrapper[4862]: I1128 11:23:43.837896 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:43 crc kubenswrapper[4862]: E1128 11:23:43.838036 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:43 crc kubenswrapper[4862]: E1128 11:23:43.838175 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:43 crc kubenswrapper[4862]: I1128 11:23:43.838408 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:43 crc kubenswrapper[4862]: E1128 11:23:43.838574 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:44 crc kubenswrapper[4862]: I1128 11:23:44.838187 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:44 crc kubenswrapper[4862]: E1128 11:23:44.838404 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:45 crc kubenswrapper[4862]: I1128 11:23:45.838403 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:45 crc kubenswrapper[4862]: I1128 11:23:45.838453 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:45 crc kubenswrapper[4862]: I1128 11:23:45.842852 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:45 crc kubenswrapper[4862]: E1128 11:23:45.842914 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:45 crc kubenswrapper[4862]: E1128 11:23:45.843054 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:45 crc kubenswrapper[4862]: E1128 11:23:45.843074 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:46 crc kubenswrapper[4862]: I1128 11:23:46.838535 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:46 crc kubenswrapper[4862]: E1128 11:23:46.838778 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:47 crc kubenswrapper[4862]: I1128 11:23:47.838156 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:47 crc kubenswrapper[4862]: I1128 11:23:47.838212 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:47 crc kubenswrapper[4862]: E1128 11:23:47.838334 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:47 crc kubenswrapper[4862]: I1128 11:23:47.838180 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:47 crc kubenswrapper[4862]: E1128 11:23:47.838532 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:47 crc kubenswrapper[4862]: E1128 11:23:47.838746 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:48 crc kubenswrapper[4862]: I1128 11:23:48.837933 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:48 crc kubenswrapper[4862]: E1128 11:23:48.838128 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:49 crc kubenswrapper[4862]: I1128 11:23:49.838332 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:49 crc kubenswrapper[4862]: I1128 11:23:49.838370 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:49 crc kubenswrapper[4862]: I1128 11:23:49.839402 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:49 crc kubenswrapper[4862]: E1128 11:23:49.839548 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:49 crc kubenswrapper[4862]: E1128 11:23:49.839755 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:49 crc kubenswrapper[4862]: E1128 11:23:49.839884 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:50 crc kubenswrapper[4862]: I1128 11:23:50.838396 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:50 crc kubenswrapper[4862]: E1128 11:23:50.838862 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:50 crc kubenswrapper[4862]: I1128 11:23:50.839184 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:23:50 crc kubenswrapper[4862]: E1128 11:23:50.839337 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ccpp6_openshift-ovn-kubernetes(9916d783-c2df-47fa-9c38-ee7537db24d4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" Nov 28 11:23:51 crc kubenswrapper[4862]: I1128 11:23:51.837828 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:51 crc kubenswrapper[4862]: I1128 11:23:51.837960 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:51 crc kubenswrapper[4862]: E1128 11:23:51.838000 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:51 crc kubenswrapper[4862]: I1128 11:23:51.838173 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:51 crc kubenswrapper[4862]: E1128 11:23:51.838282 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:51 crc kubenswrapper[4862]: E1128 11:23:51.838311 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:52 crc kubenswrapper[4862]: I1128 11:23:52.838459 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:52 crc kubenswrapper[4862]: E1128 11:23:52.840385 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:53 crc kubenswrapper[4862]: I1128 11:23:53.838670 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:53 crc kubenswrapper[4862]: I1128 11:23:53.838759 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:53 crc kubenswrapper[4862]: E1128 11:23:53.838832 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:53 crc kubenswrapper[4862]: E1128 11:23:53.839000 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:53 crc kubenswrapper[4862]: I1128 11:23:53.839315 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:53 crc kubenswrapper[4862]: E1128 11:23:53.839443 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:54 crc kubenswrapper[4862]: I1128 11:23:54.837766 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:54 crc kubenswrapper[4862]: E1128 11:23:54.837964 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:55 crc kubenswrapper[4862]: I1128 11:23:55.837791 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:55 crc kubenswrapper[4862]: I1128 11:23:55.837825 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:55 crc kubenswrapper[4862]: E1128 11:23:55.840372 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:55 crc kubenswrapper[4862]: E1128 11:23:55.840527 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:55 crc kubenswrapper[4862]: I1128 11:23:55.837887 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:55 crc kubenswrapper[4862]: E1128 11:23:55.840655 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:56 crc kubenswrapper[4862]: I1128 11:23:56.543846 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kqpd4_d6abffd5-6086-4a32-ba8d-e7c669548c5f/kube-multus/1.log" Nov 28 11:23:56 crc kubenswrapper[4862]: I1128 11:23:56.544846 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kqpd4_d6abffd5-6086-4a32-ba8d-e7c669548c5f/kube-multus/0.log" Nov 28 11:23:56 crc kubenswrapper[4862]: I1128 11:23:56.544960 4862 generic.go:334] "Generic (PLEG): container finished" podID="d6abffd5-6086-4a32-ba8d-e7c669548c5f" containerID="743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625" exitCode=1 Nov 28 11:23:56 crc kubenswrapper[4862]: I1128 11:23:56.545023 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kqpd4" event={"ID":"d6abffd5-6086-4a32-ba8d-e7c669548c5f","Type":"ContainerDied","Data":"743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625"} Nov 28 11:23:56 crc kubenswrapper[4862]: I1128 11:23:56.545154 4862 scope.go:117] "RemoveContainer" containerID="3ff163fbe2f4936f2839d9273f99de8f3a601f8650b791c065c6c3748122bfb4" Nov 28 11:23:56 crc kubenswrapper[4862]: I1128 11:23:56.545700 4862 scope.go:117] "RemoveContainer" containerID="743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625" Nov 28 11:23:56 crc kubenswrapper[4862]: E1128 11:23:56.546082 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-kqpd4_openshift-multus(d6abffd5-6086-4a32-ba8d-e7c669548c5f)\"" pod="openshift-multus/multus-kqpd4" podUID="d6abffd5-6086-4a32-ba8d-e7c669548c5f" Nov 28 11:23:56 crc kubenswrapper[4862]: I1128 11:23:56.838188 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:56 crc kubenswrapper[4862]: E1128 11:23:56.838753 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:57 crc kubenswrapper[4862]: I1128 11:23:57.551640 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kqpd4_d6abffd5-6086-4a32-ba8d-e7c669548c5f/kube-multus/1.log" Nov 28 11:23:57 crc kubenswrapper[4862]: I1128 11:23:57.837699 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:57 crc kubenswrapper[4862]: E1128 11:23:57.837856 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:57 crc kubenswrapper[4862]: I1128 11:23:57.837699 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:57 crc kubenswrapper[4862]: I1128 11:23:57.837971 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:57 crc kubenswrapper[4862]: E1128 11:23:57.838205 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:57 crc kubenswrapper[4862]: E1128 11:23:57.838300 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:23:58 crc kubenswrapper[4862]: I1128 11:23:58.837775 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:23:58 crc kubenswrapper[4862]: E1128 11:23:58.837997 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:23:59 crc kubenswrapper[4862]: I1128 11:23:59.838028 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:23:59 crc kubenswrapper[4862]: I1128 11:23:59.838166 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:23:59 crc kubenswrapper[4862]: E1128 11:23:59.838239 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:23:59 crc kubenswrapper[4862]: I1128 11:23:59.838533 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:23:59 crc kubenswrapper[4862]: E1128 11:23:59.838707 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:23:59 crc kubenswrapper[4862]: E1128 11:23:59.839149 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:00 crc kubenswrapper[4862]: I1128 11:24:00.838302 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:00 crc kubenswrapper[4862]: E1128 11:24:00.838505 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:01 crc kubenswrapper[4862]: I1128 11:24:01.838661 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:01 crc kubenswrapper[4862]: I1128 11:24:01.838695 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:01 crc kubenswrapper[4862]: E1128 11:24:01.838813 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:01 crc kubenswrapper[4862]: I1128 11:24:01.838879 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:01 crc kubenswrapper[4862]: E1128 11:24:01.839159 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:24:01 crc kubenswrapper[4862]: E1128 11:24:01.839349 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:24:02 crc kubenswrapper[4862]: E1128 11:24:02.769255 4862 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 28 11:24:02 crc kubenswrapper[4862]: I1128 11:24:02.838438 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:02 crc kubenswrapper[4862]: E1128 11:24:02.840454 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:02 crc kubenswrapper[4862]: E1128 11:24:02.954471 4862 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 11:24:03 crc kubenswrapper[4862]: I1128 11:24:03.838232 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:03 crc kubenswrapper[4862]: I1128 11:24:03.838232 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:03 crc kubenswrapper[4862]: E1128 11:24:03.838441 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:24:03 crc kubenswrapper[4862]: I1128 11:24:03.838232 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:03 crc kubenswrapper[4862]: E1128 11:24:03.838579 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:24:03 crc kubenswrapper[4862]: E1128 11:24:03.838835 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:04 crc kubenswrapper[4862]: I1128 11:24:04.839334 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:04 crc kubenswrapper[4862]: I1128 11:24:04.840357 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:24:04 crc kubenswrapper[4862]: E1128 11:24:04.841272 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:05 crc kubenswrapper[4862]: I1128 11:24:05.584036 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/3.log" Nov 28 11:24:05 crc kubenswrapper[4862]: I1128 11:24:05.587839 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerStarted","Data":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} Nov 28 11:24:05 crc kubenswrapper[4862]: I1128 11:24:05.588552 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:24:05 crc kubenswrapper[4862]: I1128 11:24:05.674039 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podStartSLOduration=103.674024779 podStartE2EDuration="1m43.674024779s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:05.671619565 +0000 UTC m=+123.204133486" watchObservedRunningTime="2025-11-28 11:24:05.674024779 +0000 UTC m=+123.206538700" Nov 28 11:24:05 crc kubenswrapper[4862]: I1128 11:24:05.837933 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:05 crc kubenswrapper[4862]: I1128 11:24:05.837937 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:05 crc kubenswrapper[4862]: I1128 11:24:05.838122 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:05 crc kubenswrapper[4862]: E1128 11:24:05.838331 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:24:05 crc kubenswrapper[4862]: E1128 11:24:05.838448 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:05 crc kubenswrapper[4862]: E1128 11:24:05.838602 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:24:05 crc kubenswrapper[4862]: I1128 11:24:05.976080 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-wl4cj"] Nov 28 11:24:05 crc kubenswrapper[4862]: I1128 11:24:05.976223 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:05 crc kubenswrapper[4862]: E1128 11:24:05.976325 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:07 crc kubenswrapper[4862]: I1128 11:24:07.837997 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:07 crc kubenswrapper[4862]: I1128 11:24:07.838040 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:07 crc kubenswrapper[4862]: I1128 11:24:07.838113 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:07 crc kubenswrapper[4862]: I1128 11:24:07.838006 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:07 crc kubenswrapper[4862]: E1128 11:24:07.838229 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:24:07 crc kubenswrapper[4862]: E1128 11:24:07.838372 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:24:07 crc kubenswrapper[4862]: E1128 11:24:07.838455 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:07 crc kubenswrapper[4862]: E1128 11:24:07.838576 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:07 crc kubenswrapper[4862]: E1128 11:24:07.955803 4862 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 11:24:09 crc kubenswrapper[4862]: I1128 11:24:09.838524 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:09 crc kubenswrapper[4862]: I1128 11:24:09.838573 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:09 crc kubenswrapper[4862]: I1128 11:24:09.838544 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:09 crc kubenswrapper[4862]: I1128 11:24:09.838703 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:09 crc kubenswrapper[4862]: E1128 11:24:09.838847 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:24:09 crc kubenswrapper[4862]: E1128 11:24:09.839057 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:09 crc kubenswrapper[4862]: E1128 11:24:09.839283 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:24:09 crc kubenswrapper[4862]: E1128 11:24:09.839359 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:10 crc kubenswrapper[4862]: I1128 11:24:10.839429 4862 scope.go:117] "RemoveContainer" containerID="743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625" Nov 28 11:24:11 crc kubenswrapper[4862]: I1128 11:24:11.612948 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kqpd4_d6abffd5-6086-4a32-ba8d-e7c669548c5f/kube-multus/1.log" Nov 28 11:24:11 crc kubenswrapper[4862]: I1128 11:24:11.613395 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kqpd4" event={"ID":"d6abffd5-6086-4a32-ba8d-e7c669548c5f","Type":"ContainerStarted","Data":"e805b51f4db68a36820f093a795585bd38dc9ad4bb444695e78ebef9a461111e"} Nov 28 11:24:11 crc kubenswrapper[4862]: I1128 11:24:11.838499 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:11 crc kubenswrapper[4862]: I1128 11:24:11.838549 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:11 crc kubenswrapper[4862]: I1128 11:24:11.838656 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:11 crc kubenswrapper[4862]: E1128 11:24:11.838694 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:24:11 crc kubenswrapper[4862]: I1128 11:24:11.838688 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:11 crc kubenswrapper[4862]: E1128 11:24:11.838747 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:24:11 crc kubenswrapper[4862]: E1128 11:24:11.838943 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:11 crc kubenswrapper[4862]: E1128 11:24:11.839174 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:12 crc kubenswrapper[4862]: E1128 11:24:12.956704 4862 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 11:24:13 crc kubenswrapper[4862]: I1128 11:24:13.838372 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:13 crc kubenswrapper[4862]: I1128 11:24:13.838476 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:13 crc kubenswrapper[4862]: I1128 11:24:13.838453 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:13 crc kubenswrapper[4862]: I1128 11:24:13.838406 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:13 crc kubenswrapper[4862]: E1128 11:24:13.838613 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:24:13 crc kubenswrapper[4862]: E1128 11:24:13.838819 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:13 crc kubenswrapper[4862]: E1128 11:24:13.839007 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:13 crc kubenswrapper[4862]: E1128 11:24:13.839210 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:24:15 crc kubenswrapper[4862]: I1128 11:24:15.838418 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:15 crc kubenswrapper[4862]: I1128 11:24:15.838482 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:15 crc kubenswrapper[4862]: E1128 11:24:15.839559 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:24:15 crc kubenswrapper[4862]: I1128 11:24:15.838538 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:15 crc kubenswrapper[4862]: E1128 11:24:15.839366 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:24:15 crc kubenswrapper[4862]: I1128 11:24:15.838511 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:15 crc kubenswrapper[4862]: E1128 11:24:15.839891 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:15 crc kubenswrapper[4862]: E1128 11:24:15.839755 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:17 crc kubenswrapper[4862]: I1128 11:24:17.595496 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:24:17 crc kubenswrapper[4862]: I1128 11:24:17.838176 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:17 crc kubenswrapper[4862]: I1128 11:24:17.838210 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:17 crc kubenswrapper[4862]: I1128 11:24:17.838264 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:17 crc kubenswrapper[4862]: E1128 11:24:17.838933 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 11:24:17 crc kubenswrapper[4862]: I1128 11:24:17.838307 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:17 crc kubenswrapper[4862]: E1128 11:24:17.839068 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 11:24:17 crc kubenswrapper[4862]: E1128 11:24:17.839360 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 11:24:17 crc kubenswrapper[4862]: E1128 11:24:17.839417 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl4cj" podUID="60f0df65-b253-4ea2-abda-dbc50f7f2eca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.371901 4862 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.430698 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8bqmb"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.431354 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.436672 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.437270 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.438568 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wt49p"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.439429 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.442764 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.442997 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.442872 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.443426 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.443598 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.443840 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.444020 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.445337 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.445840 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.445435 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.447161 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.447281 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.447519 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.447631 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.447811 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.447878 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.448005 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.448148 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.448335 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.449152 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.452316 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/336fbe7b-be62-47ab-bca9-d847a7839b99-serving-cert\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.452599 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcqhq\" (UniqueName: \"kubernetes.io/projected/336fbe7b-be62-47ab-bca9-d847a7839b99-kube-api-access-tcqhq\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.452781 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336fbe7b-be62-47ab-bca9-d847a7839b99-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.452971 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/336fbe7b-be62-47ab-bca9-d847a7839b99-config\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.453284 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336fbe7b-be62-47ab-bca9-d847a7839b99-service-ca-bundle\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.463183 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-pzshw"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.463783 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.473142 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.480362 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.480641 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.480718 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.480893 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.482018 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.482453 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.482704 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.482926 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.483121 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.483005 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qlr6s"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.483616 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.484451 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.485332 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.485590 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.492218 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8wmwz"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.492824 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.493531 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.494038 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.496492 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.502139 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhl4q"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.502465 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.502579 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.502799 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.503463 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.504034 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.504481 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.505212 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.505220 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.505505 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hw6dt"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.505567 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.505840 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.505978 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.506256 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.505857 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.506814 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.506459 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.507462 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.507629 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.507792 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.507984 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.508205 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.523675 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5hjfs"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.544190 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.544808 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.545200 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.545439 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.546197 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.546439 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.546563 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.546958 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.545553 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.545899 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.550830 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.553597 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.544334 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.554025 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7mhnb"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.555457 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.555795 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.555914 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336fbe7b-be62-47ab-bca9-d847a7839b99-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.555949 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.555968 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.555999 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd2qx\" (UniqueName: \"kubernetes.io/projected/da2a67b7-90a1-4083-b303-9c5e7a574d58-kube-api-access-kd2qx\") pod \"openshift-apiserver-operator-796bbdcf4f-lhslh\" (UID: \"da2a67b7-90a1-4083-b303-9c5e7a574d58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556029 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37b54aa2-ee51-4110-b25b-0739965fe748-serving-cert\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556052 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556054 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f85vz\" (UniqueName: \"kubernetes.io/projected/f112a896-221b-4cfc-aef3-417a60c5be62-kube-api-access-f85vz\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556393 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336fbe7b-be62-47ab-bca9-d847a7839b99-service-ca-bundle\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556439 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da2a67b7-90a1-4083-b303-9c5e7a574d58-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lhslh\" (UID: \"da2a67b7-90a1-4083-b303-9c5e7a574d58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556475 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1b858357-d43f-4e86-b212-479683104d02-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kt5hh\" (UID: \"1b858357-d43f-4e86-b212-479683104d02\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556532 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-serving-cert\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556573 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3ad0f875-1ae0-4623-807e-8e88031bbd68-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556610 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-config\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556644 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2a67b7-90a1-4083-b303-9c5e7a574d58-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lhslh\" (UID: \"da2a67b7-90a1-4083-b303-9c5e7a574d58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556676 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f112a896-221b-4cfc-aef3-417a60c5be62-etcd-client\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556718 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f112a896-221b-4cfc-aef3-417a60c5be62-encryption-config\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556751 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-config\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556783 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f112a896-221b-4cfc-aef3-417a60c5be62-audit-dir\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556815 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556868 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/336fbe7b-be62-47ab-bca9-d847a7839b99-serving-cert\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556906 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcqhq\" (UniqueName: \"kubernetes.io/projected/336fbe7b-be62-47ab-bca9-d847a7839b99-kube-api-access-tcqhq\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556942 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7212be9-9c79-49f2-8f00-db24967afe22-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-j5pxc\" (UID: \"d7212be9-9c79-49f2-8f00-db24967afe22\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.556980 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-image-import-ca\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557016 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/336fbe7b-be62-47ab-bca9-d847a7839b99-config\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557053 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbfg8\" (UniqueName: \"kubernetes.io/projected/37b54aa2-ee51-4110-b25b-0739965fe748-kube-api-access-hbfg8\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557085 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwkc6\" (UniqueName: \"kubernetes.io/projected/dda64eb0-d331-41a5-a33e-323e2438c06f-kube-api-access-wwkc6\") pod \"migrator-59844c95c7-mlwqk\" (UID: \"dda64eb0-d331-41a5-a33e-323e2438c06f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557141 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc2rw\" (UniqueName: \"kubernetes.io/projected/3ad0f875-1ae0-4623-807e-8e88031bbd68-kube-api-access-qc2rw\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557180 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7212be9-9c79-49f2-8f00-db24967afe22-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-j5pxc\" (UID: \"d7212be9-9c79-49f2-8f00-db24967afe22\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557210 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f112a896-221b-4cfc-aef3-417a60c5be62-node-pullsecrets\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557244 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-audit\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557275 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9gvl\" (UniqueName: \"kubernetes.io/projected/1b858357-d43f-4e86-b212-479683104d02-kube-api-access-t9gvl\") pod \"openshift-config-operator-7777fb866f-kt5hh\" (UID: \"1b858357-d43f-4e86-b212-479683104d02\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557307 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37b54aa2-ee51-4110-b25b-0739965fe748-trusted-ca\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557336 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-etcd-serving-ca\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557386 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b54aa2-ee51-4110-b25b-0739965fe748-config\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557436 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ad0f875-1ae0-4623-807e-8e88031bbd68-config\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557437 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336fbe7b-be62-47ab-bca9-d847a7839b99-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.555879 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557623 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b858357-d43f-4e86-b212-479683104d02-serving-cert\") pod \"openshift-config-operator-7777fb866f-kt5hh\" (UID: \"1b858357-d43f-4e86-b212-479683104d02\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.557959 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f112a896-221b-4cfc-aef3-417a60c5be62-serving-cert\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.558143 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.558315 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336fbe7b-be62-47ab-bca9-d847a7839b99-service-ca-bundle\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.558981 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/336fbe7b-be62-47ab-bca9-d847a7839b99-config\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.559614 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-client-ca\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.558248 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.559680 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3ad0f875-1ae0-4623-807e-8e88031bbd68-images\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.559698 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7212be9-9c79-49f2-8f00-db24967afe22-config\") pod \"kube-apiserver-operator-766d6c64bb-j5pxc\" (UID: \"d7212be9-9c79-49f2-8f00-db24967afe22\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.558458 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.559720 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqtx8\" (UniqueName: \"kubernetes.io/projected/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-kube-api-access-qqtx8\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.558488 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.558720 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.560188 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.560693 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-45tfc"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.560914 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.561073 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-45tfc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.561282 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.561831 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.562227 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.562639 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.564182 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.565735 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.569223 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.569223 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.569920 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.579441 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g5fhw"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.580152 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.582573 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qlsvq"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.584045 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.586131 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.586597 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/336fbe7b-be62-47ab-bca9-d847a7839b99-serving-cert\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.586929 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.587755 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.587854 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.588390 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.588562 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.588728 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.588897 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.589063 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.589190 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.589295 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.589384 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.589480 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.589566 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.589807 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.589909 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.590006 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.590175 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.590217 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.590313 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.590409 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.590509 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.590616 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.590727 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.590856 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.591084 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.592885 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.596517 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.598818 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.598990 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-2bhn7"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.600028 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.608206 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcqhq\" (UniqueName: \"kubernetes.io/projected/336fbe7b-be62-47ab-bca9-d847a7839b99-kube-api-access-tcqhq\") pod \"authentication-operator-69f744f599-8bqmb\" (UID: \"336fbe7b-be62-47ab-bca9-d847a7839b99\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.608286 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.609783 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.616446 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.621717 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.624109 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.635686 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.651956 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.654987 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.655332 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.655694 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.655788 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.655972 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.655689 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660639 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-etcd-serving-ca\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660683 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37b54aa2-ee51-4110-b25b-0739965fe748-trusted-ca\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660701 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b54aa2-ee51-4110-b25b-0739965fe748-config\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660728 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d275feb5-1839-4942-9330-c4f5c47cc0fe-config\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660746 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ad0f875-1ae0-4623-807e-8e88031bbd68-config\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660763 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b858357-d43f-4e86-b212-479683104d02-serving-cert\") pod \"openshift-config-operator-7777fb866f-kt5hh\" (UID: \"1b858357-d43f-4e86-b212-479683104d02\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660779 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-oauth-config\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660797 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-config\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660815 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f112a896-221b-4cfc-aef3-417a60c5be62-serving-cert\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660831 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-certificates\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660848 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-oauth-serving-cert\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660872 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3ad0f875-1ae0-4623-807e-8e88031bbd68-images\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660888 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-client-ca\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660906 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7212be9-9c79-49f2-8f00-db24967afe22-config\") pod \"kube-apiserver-operator-766d6c64bb-j5pxc\" (UID: \"d7212be9-9c79-49f2-8f00-db24967afe22\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660922 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqtx8\" (UniqueName: \"kubernetes.io/projected/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-kube-api-access-qqtx8\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660939 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d275feb5-1839-4942-9330-c4f5c47cc0fe-etcd-ca\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660960 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-tls\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660976 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wcjf\" (UniqueName: \"kubernetes.io/projected/d275feb5-1839-4942-9330-c4f5c47cc0fe-kube-api-access-2wcjf\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.660993 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661012 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd2qx\" (UniqueName: \"kubernetes.io/projected/da2a67b7-90a1-4083-b303-9c5e7a574d58-kube-api-access-kd2qx\") pod \"openshift-apiserver-operator-796bbdcf4f-lhslh\" (UID: \"da2a67b7-90a1-4083-b303-9c5e7a574d58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661031 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6e143937-4cf3-4fa0-aa09-381d8bf99f16-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661049 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-trusted-ca-bundle\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661067 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f85vz\" (UniqueName: \"kubernetes.io/projected/f112a896-221b-4cfc-aef3-417a60c5be62-kube-api-access-f85vz\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661101 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661121 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-service-ca\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661138 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37b54aa2-ee51-4110-b25b-0739965fe748-serving-cert\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661156 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da2a67b7-90a1-4083-b303-9c5e7a574d58-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lhslh\" (UID: \"da2a67b7-90a1-4083-b303-9c5e7a574d58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661173 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1b858357-d43f-4e86-b212-479683104d02-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kt5hh\" (UID: \"1b858357-d43f-4e86-b212-479683104d02\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661192 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6e143937-4cf3-4fa0-aa09-381d8bf99f16-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661209 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d275feb5-1839-4942-9330-c4f5c47cc0fe-serving-cert\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661224 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-serving-cert\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661241 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-serving-cert\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661268 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3ad0f875-1ae0-4623-807e-8e88031bbd68-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661288 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-config\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661306 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-bound-sa-token\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661323 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2a67b7-90a1-4083-b303-9c5e7a574d58-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lhslh\" (UID: \"da2a67b7-90a1-4083-b303-9c5e7a574d58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661340 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f112a896-221b-4cfc-aef3-417a60c5be62-etcd-client\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661358 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f112a896-221b-4cfc-aef3-417a60c5be62-encryption-config\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661372 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-config\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661389 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661408 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8fwm\" (UniqueName: \"kubernetes.io/projected/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-kube-api-access-k8fwm\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661424 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f112a896-221b-4cfc-aef3-417a60c5be62-audit-dir\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661448 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7212be9-9c79-49f2-8f00-db24967afe22-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-j5pxc\" (UID: \"d7212be9-9c79-49f2-8f00-db24967afe22\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661464 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7plxg\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-kube-api-access-7plxg\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661481 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-trusted-ca\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661499 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-image-import-ca\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661514 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d275feb5-1839-4942-9330-c4f5c47cc0fe-etcd-client\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661531 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbfg8\" (UniqueName: \"kubernetes.io/projected/37b54aa2-ee51-4110-b25b-0739965fe748-kube-api-access-hbfg8\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661549 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwkc6\" (UniqueName: \"kubernetes.io/projected/dda64eb0-d331-41a5-a33e-323e2438c06f-kube-api-access-wwkc6\") pod \"migrator-59844c95c7-mlwqk\" (UID: \"dda64eb0-d331-41a5-a33e-323e2438c06f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661563 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc2rw\" (UniqueName: \"kubernetes.io/projected/3ad0f875-1ae0-4623-807e-8e88031bbd68-kube-api-access-qc2rw\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661580 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7212be9-9c79-49f2-8f00-db24967afe22-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-j5pxc\" (UID: \"d7212be9-9c79-49f2-8f00-db24967afe22\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661595 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9gvl\" (UniqueName: \"kubernetes.io/projected/1b858357-d43f-4e86-b212-479683104d02-kube-api-access-t9gvl\") pod \"openshift-config-operator-7777fb866f-kt5hh\" (UID: \"1b858357-d43f-4e86-b212-479683104d02\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661611 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d275feb5-1839-4942-9330-c4f5c47cc0fe-etcd-service-ca\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661626 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f112a896-221b-4cfc-aef3-417a60c5be62-node-pullsecrets\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.661652 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-audit\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.662301 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-audit\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.662713 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-etcd-serving-ca\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.663796 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1b858357-d43f-4e86-b212-479683104d02-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kt5hh\" (UID: \"1b858357-d43f-4e86-b212-479683104d02\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.664339 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b54aa2-ee51-4110-b25b-0739965fe748-config\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.664998 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ad0f875-1ae0-4623-807e-8e88031bbd68-config\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.665072 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-s926b"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.665571 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3ad0f875-1ae0-4623-807e-8e88031bbd68-images\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.665842 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.666410 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.666430 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-client-ca\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: E1128 11:24:18.667767 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.167751025 +0000 UTC m=+136.700264946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.668129 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7212be9-9c79-49f2-8f00-db24967afe22-config\") pod \"kube-apiserver-operator-766d6c64bb-j5pxc\" (UID: \"d7212be9-9c79-49f2-8f00-db24967afe22\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.668772 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f112a896-221b-4cfc-aef3-417a60c5be62-audit-dir\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.669259 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37b54aa2-ee51-4110-b25b-0739965fe748-trusted-ca\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.669508 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.669920 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.670425 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-serving-cert\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.670451 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-s86df"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.670522 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.670624 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-image-import-ca\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.670936 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.671072 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.671323 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2a67b7-90a1-4083-b303-9c5e7a574d58-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lhslh\" (UID: \"da2a67b7-90a1-4083-b303-9c5e7a574d58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.672132 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37b54aa2-ee51-4110-b25b-0739965fe748-serving-cert\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.672706 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f112a896-221b-4cfc-aef3-417a60c5be62-config\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.672765 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f112a896-221b-4cfc-aef3-417a60c5be62-node-pullsecrets\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.673370 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f112a896-221b-4cfc-aef3-417a60c5be62-etcd-client\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.673439 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nbzb4"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.673994 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.674410 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.674483 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.674570 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.674683 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.675383 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.675557 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.675698 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3ad0f875-1ae0-4623-807e-8e88031bbd68-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.676682 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.676884 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f112a896-221b-4cfc-aef3-417a60c5be62-encryption-config\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.676906 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b858357-d43f-4e86-b212-479683104d02-serving-cert\") pod \"openshift-config-operator-7777fb866f-kt5hh\" (UID: \"1b858357-d43f-4e86-b212-479683104d02\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.676933 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.677403 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.677572 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da2a67b7-90a1-4083-b303-9c5e7a574d58-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lhslh\" (UID: \"da2a67b7-90a1-4083-b303-9c5e7a574d58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.678045 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7212be9-9c79-49f2-8f00-db24967afe22-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-j5pxc\" (UID: \"d7212be9-9c79-49f2-8f00-db24967afe22\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.678122 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.679623 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.680277 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f112a896-221b-4cfc-aef3-417a60c5be62-serving-cert\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.689874 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q7bbb"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.690842 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-config\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.690898 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.693807 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8bqmb"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.695980 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.696642 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.697605 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.697833 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.699079 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-nqpc6"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.699592 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.700643 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.703158 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pzshw"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.706068 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.709017 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qlr6s"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.711321 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.714431 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.714460 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.715787 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-s926b"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.716974 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g5fhw"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.718455 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.722470 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.726267 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7mhnb"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.731449 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhl4q"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.732148 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.733166 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.733542 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.734280 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.736643 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8wmwz"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.737747 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nbzb4"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.738791 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.739891 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5hjfs"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.741118 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-45tfc"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.742689 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.745327 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-s86df"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.746970 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.748015 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hw6dt"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.749070 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.750156 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.751196 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.752329 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wt49p"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.753702 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.754603 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-l8mhs"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.755935 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-kpm6d"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.756274 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.756653 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qlsvq"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.756677 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.756758 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.757581 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.758485 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.759530 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.761702 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q7bbb"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762084 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:18 crc kubenswrapper[4862]: E1128 11:24:18.762194 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.262174998 +0000 UTC m=+136.794688919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762267 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7plxg\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-kube-api-access-7plxg\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762298 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-machine-approver-tls\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762319 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-srv-cert\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762339 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-policies\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762360 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-trusted-ca\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762380 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67f7715f-3bb2-4b68-9f0c-fda6aee68678-serving-cert\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762401 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tx94r\" (UID: \"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762555 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tx94r\" (UID: \"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762712 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762735 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4wk4\" (UniqueName: \"kubernetes.io/projected/c34ee8a3-cf6f-4d17-9638-8950c56b9fc8-kube-api-access-q4wk4\") pod \"multus-admission-controller-857f4d67dd-qlsvq\" (UID: \"c34ee8a3-cf6f-4d17-9638-8950c56b9fc8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762860 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.762934 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d275feb5-1839-4942-9330-c4f5c47cc0fe-etcd-service-ca\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763007 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763104 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfklc\" (UniqueName: \"kubernetes.io/projected/3b577ecb-1896-4699-bd7e-33929f2cf9a3-kube-api-access-cfklc\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763179 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mth8\" (UniqueName: \"kubernetes.io/projected/5c0e4c02-c077-4c5e-bc39-0601250102b3-kube-api-access-7mth8\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763270 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce5a45b8-754f-46aa-a0eb-8325aa5a3868-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gvpct\" (UID: \"ce5a45b8-754f-46aa-a0eb-8325aa5a3868\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763360 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-config\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763443 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x72m\" (UniqueName: \"kubernetes.io/projected/9b3ee9cc-2155-4e40-a8e3-6ed95db6962b-kube-api-access-4x72m\") pod \"cluster-samples-operator-665b6dd947-9f6sc\" (UID: \"9b3ee9cc-2155-4e40-a8e3-6ed95db6962b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763515 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-trusted-ca\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763597 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763687 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/170cfb20-5cf0-4050-944b-6b17b0e2decc-metrics-tls\") pod \"dns-operator-744455d44c-7mhnb\" (UID: \"170cfb20-5cf0-4050-944b-6b17b0e2decc\") " pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763763 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763834 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee1f3625-a084-4e0d-99af-627305b0f8f2-serving-cert\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763905 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wcjf\" (UniqueName: \"kubernetes.io/projected/d275feb5-1839-4942-9330-c4f5c47cc0fe-kube-api-access-2wcjf\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.763785 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kpm6d"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764009 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764047 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d275feb5-1839-4942-9330-c4f5c47cc0fe-etcd-service-ca\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764077 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-audit-policies\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764130 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-serving-cert\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764178 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a1a2178-787b-4af4-a15f-225ec25b4216-config-volume\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764584 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6e143937-4cf3-4fa0-aa09-381d8bf99f16-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764628 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-config\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764652 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f40b08d-cf89-4ff6-a766-a38aab93f8dd-config\") pod \"kube-controller-manager-operator-78b949d7b-xw6qk\" (UID: \"3f40b08d-cf89-4ff6-a766-a38aab93f8dd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764709 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6lq7\" (UniqueName: \"kubernetes.io/projected/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-kube-api-access-l6lq7\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764736 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce5a45b8-754f-46aa-a0eb-8325aa5a3868-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gvpct\" (UID: \"ce5a45b8-754f-46aa-a0eb-8325aa5a3868\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764767 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6e143937-4cf3-4fa0-aa09-381d8bf99f16-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764796 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-etcd-client\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764825 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c0e4c02-c077-4c5e-bc39-0601250102b3-trusted-ca\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764888 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764921 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765060 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6e143937-4cf3-4fa0-aa09-381d8bf99f16-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765250 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-l8mhs"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.764927 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djt7v\" (UniqueName: \"kubernetes.io/projected/56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f-kube-api-access-djt7v\") pod \"kube-storage-version-migrator-operator-b67b599dd-tx94r\" (UID: \"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765412 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765443 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f45f\" (UniqueName: \"kubernetes.io/projected/a4038318-be77-4a50-a6f2-611308bda6e7-kube-api-access-6f45f\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765478 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce5a45b8-754f-46aa-a0eb-8325aa5a3868-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gvpct\" (UID: \"ce5a45b8-754f-46aa-a0eb-8325aa5a3868\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765506 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765555 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-encryption-config\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765668 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5c0e4c02-c077-4c5e-bc39-0601250102b3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765722 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d275feb5-1839-4942-9330-c4f5c47cc0fe-etcd-client\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765762 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-config\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765792 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765831 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67f7715f-3bb2-4b68-9f0c-fda6aee68678-config\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765861 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a1a2178-787b-4af4-a15f-225ec25b4216-secret-volume\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765893 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-client-ca\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.765958 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-config\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766009 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d275feb5-1839-4942-9330-c4f5c47cc0fe-config\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766044 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-audit-dir\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766078 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll696\" (UniqueName: \"kubernetes.io/projected/3a1a2178-787b-4af4-a15f-225ec25b4216-kube-api-access-ll696\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766140 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-oauth-config\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766182 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f40b08d-cf89-4ff6-a766-a38aab93f8dd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xw6qk\" (UID: \"3f40b08d-cf89-4ff6-a766-a38aab93f8dd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766214 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766251 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-certificates\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766280 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-oauth-serving-cert\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766311 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54gdk\" (UniqueName: \"kubernetes.io/projected/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-kube-api-access-54gdk\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766342 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b3ee9cc-2155-4e40-a8e3-6ed95db6962b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9f6sc\" (UID: \"9b3ee9cc-2155-4e40-a8e3-6ed95db6962b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766374 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qxfz\" (UniqueName: \"kubernetes.io/projected/170cfb20-5cf0-4050-944b-6b17b0e2decc-kube-api-access-5qxfz\") pod \"dns-operator-744455d44c-7mhnb\" (UID: \"170cfb20-5cf0-4050-944b-6b17b0e2decc\") " pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766403 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766448 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d275feb5-1839-4942-9330-c4f5c47cc0fe-etcd-ca\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766482 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766514 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f70be661-2a14-4ba9-a271-ce98998cf14c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wndxn\" (UID: \"f70be661-2a14-4ba9-a271-ce98998cf14c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766544 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljvzf\" (UniqueName: \"kubernetes.io/projected/f70be661-2a14-4ba9-a271-ce98998cf14c-kube-api-access-ljvzf\") pod \"package-server-manager-789f6589d5-wndxn\" (UID: \"f70be661-2a14-4ba9-a271-ce98998cf14c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766571 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766604 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-tls\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766648 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c34ee8a3-cf6f-4d17-9638-8950c56b9fc8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qlsvq\" (UID: \"c34ee8a3-cf6f-4d17-9638-8950c56b9fc8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766701 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-trusted-ca-bundle\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766775 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766808 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-service-ca\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766838 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f40b08d-cf89-4ff6-a766-a38aab93f8dd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xw6qk\" (UID: \"3f40b08d-cf89-4ff6-a766-a38aab93f8dd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766875 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766905 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d275feb5-1839-4942-9330-c4f5c47cc0fe-serving-cert\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766932 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-288ph\" (UniqueName: \"kubernetes.io/projected/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-kube-api-access-288ph\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766953 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-dir\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766979 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-d5qm9"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.767568 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-d5qm9"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.767649 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-d5qm9" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.766983 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-serving-cert\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.767945 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.767987 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxxpb\" (UniqueName: \"kubernetes.io/projected/67f7715f-3bb2-4b68-9f0c-fda6aee68678-kube-api-access-fxxpb\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.768014 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.768039 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-bound-sa-token\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.768066 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5c0e4c02-c077-4c5e-bc39-0601250102b3-metrics-tls\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.768100 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-auth-proxy-config\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.768124 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sscpn\" (UniqueName: \"kubernetes.io/projected/ee1f3625-a084-4e0d-99af-627305b0f8f2-kube-api-access-sscpn\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.768162 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8fwm\" (UniqueName: \"kubernetes.io/projected/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-kube-api-access-k8fwm\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.768180 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psxr8\" (UniqueName: \"kubernetes.io/projected/faca5cde-65a6-429e-816c-a97fa57c5bf1-kube-api-access-psxr8\") pod \"downloads-7954f5f757-45tfc\" (UID: \"faca5cde-65a6-429e-816c-a97fa57c5bf1\") " pod="openshift-console/downloads-7954f5f757-45tfc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.768448 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d275feb5-1839-4942-9330-c4f5c47cc0fe-config\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.768601 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6e143937-4cf3-4fa0-aa09-381d8bf99f16-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: E1128 11:24:18.769154 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.269132719 +0000 UTC m=+136.801646700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.769678 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-service-ca\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.770460 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-oauth-serving-cert\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.770507 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d275feb5-1839-4942-9330-c4f5c47cc0fe-etcd-ca\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.770599 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-serving-cert\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.770884 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-certificates\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.771270 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d275feb5-1839-4942-9330-c4f5c47cc0fe-etcd-client\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.771558 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-trusted-ca-bundle\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.771817 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-tls\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.772256 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-oauth-config\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.776859 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.781019 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d275feb5-1839-4942-9330-c4f5c47cc0fe-serving-cert\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.793564 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.813599 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.834489 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.854322 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.869478 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.869889 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tx94r\" (UID: \"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.869971 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4wk4\" (UniqueName: \"kubernetes.io/projected/c34ee8a3-cf6f-4d17-9638-8950c56b9fc8-kube-api-access-q4wk4\") pod \"multus-admission-controller-857f4d67dd-qlsvq\" (UID: \"c34ee8a3-cf6f-4d17-9638-8950c56b9fc8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870053 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870121 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870163 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfklc\" (UniqueName: \"kubernetes.io/projected/3b577ecb-1896-4699-bd7e-33929f2cf9a3-kube-api-access-cfklc\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870219 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mth8\" (UniqueName: \"kubernetes.io/projected/5c0e4c02-c077-4c5e-bc39-0601250102b3-kube-api-access-7mth8\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870249 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce5a45b8-754f-46aa-a0eb-8325aa5a3868-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gvpct\" (UID: \"ce5a45b8-754f-46aa-a0eb-8325aa5a3868\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870300 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x72m\" (UniqueName: \"kubernetes.io/projected/9b3ee9cc-2155-4e40-a8e3-6ed95db6962b-kube-api-access-4x72m\") pod \"cluster-samples-operator-665b6dd947-9f6sc\" (UID: \"9b3ee9cc-2155-4e40-a8e3-6ed95db6962b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870365 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870392 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870415 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/170cfb20-5cf0-4050-944b-6b17b0e2decc-metrics-tls\") pod \"dns-operator-744455d44c-7mhnb\" (UID: \"170cfb20-5cf0-4050-944b-6b17b0e2decc\") " pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870494 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee1f3625-a084-4e0d-99af-627305b0f8f2-serving-cert\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870566 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870590 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-audit-policies\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870646 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-serving-cert\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.870669 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a1a2178-787b-4af4-a15f-225ec25b4216-config-volume\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.871471 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f40b08d-cf89-4ff6-a766-a38aab93f8dd-config\") pod \"kube-controller-manager-operator-78b949d7b-xw6qk\" (UID: \"3f40b08d-cf89-4ff6-a766-a38aab93f8dd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.871503 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6lq7\" (UniqueName: \"kubernetes.io/projected/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-kube-api-access-l6lq7\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.871672 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce5a45b8-754f-46aa-a0eb-8325aa5a3868-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gvpct\" (UID: \"ce5a45b8-754f-46aa-a0eb-8325aa5a3868\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.871818 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-etcd-client\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.871843 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c0e4c02-c077-4c5e-bc39-0601250102b3-trusted-ca\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.871982 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djt7v\" (UniqueName: \"kubernetes.io/projected/56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f-kube-api-access-djt7v\") pod \"kube-storage-version-migrator-operator-b67b599dd-tx94r\" (UID: \"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.872014 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.872182 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.872471 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f45f\" (UniqueName: \"kubernetes.io/projected/a4038318-be77-4a50-a6f2-611308bda6e7-kube-api-access-6f45f\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.872615 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce5a45b8-754f-46aa-a0eb-8325aa5a3868-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gvpct\" (UID: \"ce5a45b8-754f-46aa-a0eb-8325aa5a3868\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.872641 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.872785 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-encryption-config\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.872966 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5c0e4c02-c077-4c5e-bc39-0601250102b3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.872995 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-config\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873113 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f40b08d-cf89-4ff6-a766-a38aab93f8dd-config\") pod \"kube-controller-manager-operator-78b949d7b-xw6qk\" (UID: \"3f40b08d-cf89-4ff6-a766-a38aab93f8dd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873128 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873168 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67f7715f-3bb2-4b68-9f0c-fda6aee68678-config\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873317 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a1a2178-787b-4af4-a15f-225ec25b4216-secret-volume\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873481 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-client-ca\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873509 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-config\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873669 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll696\" (UniqueName: \"kubernetes.io/projected/3a1a2178-787b-4af4-a15f-225ec25b4216-kube-api-access-ll696\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873809 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-audit-dir\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873839 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f40b08d-cf89-4ff6-a766-a38aab93f8dd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xw6qk\" (UID: \"3f40b08d-cf89-4ff6-a766-a38aab93f8dd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.873978 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.874006 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54gdk\" (UniqueName: \"kubernetes.io/projected/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-kube-api-access-54gdk\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.874137 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b3ee9cc-2155-4e40-a8e3-6ed95db6962b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9f6sc\" (UID: \"9b3ee9cc-2155-4e40-a8e3-6ed95db6962b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" Nov 28 11:24:18 crc kubenswrapper[4862]: E1128 11:24:18.874174 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.374154448 +0000 UTC m=+136.906668369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.874202 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qxfz\" (UniqueName: \"kubernetes.io/projected/170cfb20-5cf0-4050-944b-6b17b0e2decc-kube-api-access-5qxfz\") pod \"dns-operator-744455d44c-7mhnb\" (UID: \"170cfb20-5cf0-4050-944b-6b17b0e2decc\") " pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.874740 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.875239 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-audit-policies\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.875595 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-client-ca\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.876424 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.876713 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-config\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.876959 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee1f3625-a084-4e0d-99af-627305b0f8f2-serving-cert\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.877084 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-audit-dir\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.877687 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-serving-cert\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.877796 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.877960 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878219 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878251 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878282 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878304 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f70be661-2a14-4ba9-a271-ce98998cf14c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wndxn\" (UID: \"f70be661-2a14-4ba9-a271-ce98998cf14c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878324 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljvzf\" (UniqueName: \"kubernetes.io/projected/f70be661-2a14-4ba9-a271-ce98998cf14c-kube-api-access-ljvzf\") pod \"package-server-manager-789f6589d5-wndxn\" (UID: \"f70be661-2a14-4ba9-a271-ce98998cf14c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878341 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878360 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c34ee8a3-cf6f-4d17-9638-8950c56b9fc8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qlsvq\" (UID: \"c34ee8a3-cf6f-4d17-9638-8950c56b9fc8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878391 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878410 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f40b08d-cf89-4ff6-a766-a38aab93f8dd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xw6qk\" (UID: \"3f40b08d-cf89-4ff6-a766-a38aab93f8dd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878438 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878459 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-288ph\" (UniqueName: \"kubernetes.io/projected/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-kube-api-access-288ph\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878478 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-dir\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878497 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxxpb\" (UniqueName: \"kubernetes.io/projected/67f7715f-3bb2-4b68-9f0c-fda6aee68678-kube-api-access-fxxpb\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878529 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878548 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878572 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5c0e4c02-c077-4c5e-bc39-0601250102b3-metrics-tls\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878589 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-auth-proxy-config\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878607 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sscpn\" (UniqueName: \"kubernetes.io/projected/ee1f3625-a084-4e0d-99af-627305b0f8f2-kube-api-access-sscpn\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878632 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psxr8\" (UniqueName: \"kubernetes.io/projected/faca5cde-65a6-429e-816c-a97fa57c5bf1-kube-api-access-psxr8\") pod \"downloads-7954f5f757-45tfc\" (UID: \"faca5cde-65a6-429e-816c-a97fa57c5bf1\") " pod="openshift-console/downloads-7954f5f757-45tfc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878656 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-policies\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878680 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-machine-approver-tls\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878696 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-srv-cert\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878714 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67f7715f-3bb2-4b68-9f0c-fda6aee68678-serving-cert\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.878733 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tx94r\" (UID: \"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.879167 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f40b08d-cf89-4ff6-a766-a38aab93f8dd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xw6qk\" (UID: \"3f40b08d-cf89-4ff6-a766-a38aab93f8dd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.879577 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-encryption-config\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.880327 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.880917 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-etcd-client\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.881055 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-dir\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: E1128 11:24:18.881265 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.381249024 +0000 UTC m=+136.913762945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.882149 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-policies\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.884538 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b3ee9cc-2155-4e40-a8e3-6ed95db6962b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9f6sc\" (UID: \"9b3ee9cc-2155-4e40-a8e3-6ed95db6962b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.886300 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/170cfb20-5cf0-4050-944b-6b17b0e2decc-metrics-tls\") pod \"dns-operator-744455d44c-7mhnb\" (UID: \"170cfb20-5cf0-4050-944b-6b17b0e2decc\") " pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.892754 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.892967 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.893341 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.894307 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.894845 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.895775 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.904115 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.904245 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.913464 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.915215 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.934232 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.945473 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c34ee8a3-cf6f-4d17-9638-8950c56b9fc8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qlsvq\" (UID: \"c34ee8a3-cf6f-4d17-9638-8950c56b9fc8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.954659 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8bqmb"] Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.954953 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.972994 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.980256 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:18 crc kubenswrapper[4862]: E1128 11:24:18.980424 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.480397437 +0000 UTC m=+137.012911368 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.980569 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:18 crc kubenswrapper[4862]: E1128 11:24:18.980951 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.480940394 +0000 UTC m=+137.013454325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:18 crc kubenswrapper[4862]: I1128 11:24:18.994244 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.014021 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.034233 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.054312 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.076597 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.082415 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.082551 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.582531025 +0000 UTC m=+137.115044946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.082709 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.083047 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.583036651 +0000 UTC m=+137.115550572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.093159 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.133167 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.145222 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tx94r\" (UID: \"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.155126 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.174936 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.183988 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.184121 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.684074895 +0000 UTC m=+137.216588826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.184569 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.185294 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.685261273 +0000 UTC m=+137.217775224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.196241 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.202552 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tx94r\" (UID: \"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.214533 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.233420 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.255914 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.275215 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.286434 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.286677 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.786636817 +0000 UTC m=+137.319150778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.287133 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.287887 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.787856085 +0000 UTC m=+137.320370046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.294088 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.314663 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.334907 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.345583 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5c0e4c02-c077-4c5e-bc39-0601250102b3-metrics-tls\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.367030 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.374830 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.376172 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c0e4c02-c077-4c5e-bc39-0601250102b3-trusted-ca\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.388706 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.389007 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.888959801 +0000 UTC m=+137.421473752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.389353 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.389895 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.88987814 +0000 UTC m=+137.422392101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.393284 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.413868 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.415906 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce5a45b8-754f-46aa-a0eb-8325aa5a3868-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gvpct\" (UID: \"ce5a45b8-754f-46aa-a0eb-8325aa5a3868\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.436144 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.454072 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.474937 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.490635 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.490899 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.990861372 +0000 UTC m=+137.523375293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.491389 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.492072 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:19.992044479 +0000 UTC m=+137.524558410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.494407 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.515011 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.535284 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.549694 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce5a45b8-754f-46aa-a0eb-8325aa5a3868-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gvpct\" (UID: \"ce5a45b8-754f-46aa-a0eb-8325aa5a3868\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.583060 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqtx8\" (UniqueName: \"kubernetes.io/projected/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-kube-api-access-qqtx8\") pod \"controller-manager-879f6c89f-bhl4q\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.592666 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.593190 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.093146315 +0000 UTC m=+137.625660236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.614907 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.636426 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f85vz\" (UniqueName: \"kubernetes.io/projected/f112a896-221b-4cfc-aef3-417a60c5be62-kube-api-access-f85vz\") pod \"apiserver-76f77b778f-wt49p\" (UID: \"f112a896-221b-4cfc-aef3-417a60c5be62\") " pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.647490 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.651678 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd2qx\" (UniqueName: \"kubernetes.io/projected/da2a67b7-90a1-4083-b303-9c5e7a574d58-kube-api-access-kd2qx\") pod \"openshift-apiserver-operator-796bbdcf4f-lhslh\" (UID: \"da2a67b7-90a1-4083-b303-9c5e7a574d58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.653664 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" event={"ID":"336fbe7b-be62-47ab-bca9-d847a7839b99","Type":"ContainerStarted","Data":"4797f3c9d49e7d04f4ca2e6b0d821d9db1a987e41959ecee2fd4f9e1b57cebe6"} Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.653713 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" event={"ID":"336fbe7b-be62-47ab-bca9-d847a7839b99","Type":"ContainerStarted","Data":"6810b36ed075f8d8a418e2bfa863325d1b29ab53b6bac1f5b41a9b0a033d0aa8"} Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.672233 4862 request.go:700] Waited for 1.003342599s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/serviceaccounts/kube-apiserver-operator/token Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.674671 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbfg8\" (UniqueName: \"kubernetes.io/projected/37b54aa2-ee51-4110-b25b-0739965fe748-kube-api-access-hbfg8\") pod \"console-operator-58897d9998-8wmwz\" (UID: \"37b54aa2-ee51-4110-b25b-0739965fe748\") " pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.682064 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.694781 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.695886 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.195863952 +0000 UTC m=+137.728377893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.700979 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7212be9-9c79-49f2-8f00-db24967afe22-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-j5pxc\" (UID: \"d7212be9-9c79-49f2-8f00-db24967afe22\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.710429 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.730119 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc2rw\" (UniqueName: \"kubernetes.io/projected/3ad0f875-1ae0-4623-807e-8e88031bbd68-kube-api-access-qc2rw\") pod \"machine-api-operator-5694c8668f-qlr6s\" (UID: \"3ad0f875-1ae0-4623-807e-8e88031bbd68\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.731285 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwkc6\" (UniqueName: \"kubernetes.io/projected/dda64eb0-d331-41a5-a33e-323e2438c06f-kube-api-access-wwkc6\") pod \"migrator-59844c95c7-mlwqk\" (UID: \"dda64eb0-d331-41a5-a33e-323e2438c06f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.734215 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.758190 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.765044 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.774428 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.793534 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.796788 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.797037 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.297009098 +0000 UTC m=+137.829523019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.797327 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.797661 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.297651109 +0000 UTC m=+137.830165030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.813519 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.824821 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.828652 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-machine-approver-tls\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.831775 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.834901 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.837630 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.837697 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.837704 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.837880 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.841976 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-auth-proxy-config\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.859690 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.861343 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh"] Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.867495 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.867716 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-config\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.874546 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.874976 4862 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.874988 4862 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.875057 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-operator-metrics podName:3b577ecb-1896-4699-bd7e-33929f2cf9a3 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.37503545 +0000 UTC m=+137.907549371 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-operator-metrics") pod "marketplace-operator-79b997595-nbzb4" (UID: "3b577ecb-1896-4699-bd7e-33929f2cf9a3") : failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.875079 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1a2178-787b-4af4-a15f-225ec25b4216-secret-volume podName:3a1a2178-787b-4af4-a15f-225ec25b4216 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.375069681 +0000 UTC m=+137.907583602 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/3a1a2178-787b-4af4-a15f-225ec25b4216-secret-volume") pod "collect-profiles-29405475-qkjm5" (UID: "3a1a2178-787b-4af4-a15f-225ec25b4216") : failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.875130 4862 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.875302 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/67f7715f-3bb2-4b68-9f0c-fda6aee68678-config podName:67f7715f-3bb2-4b68-9f0c-fda6aee68678 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.375251877 +0000 UTC m=+137.907765798 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/67f7715f-3bb2-4b68-9f0c-fda6aee68678-config") pod "service-ca-operator-777779d784-s86df" (UID: "67f7715f-3bb2-4b68-9f0c-fda6aee68678") : failed to sync configmap cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.877475 4862 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.877591 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3a1a2178-787b-4af4-a15f-225ec25b4216-config-volume podName:3a1a2178-787b-4af4-a15f-225ec25b4216 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.37755158 +0000 UTC m=+137.910065501 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/3a1a2178-787b-4af4-a15f-225ec25b4216-config-volume") pod "collect-profiles-29405475-qkjm5" (UID: "3a1a2178-787b-4af4-a15f-225ec25b4216") : failed to sync configmap cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.877627 4862 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.877668 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-trusted-ca podName:3b577ecb-1896-4699-bd7e-33929f2cf9a3 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.377661183 +0000 UTC m=+137.910175104 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-trusted-ca") pod "marketplace-operator-79b997595-nbzb4" (UID: "3b577ecb-1896-4699-bd7e-33929f2cf9a3") : failed to sync configmap cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.881892 4862 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.881943 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-profile-collector-cert podName:a4038318-be77-4a50-a6f2-611308bda6e7 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.381932139 +0000 UTC m=+137.914446050 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-profile-collector-cert") pod "olm-operator-6b444d44fb-7jpvc" (UID: "a4038318-be77-4a50-a6f2-611308bda6e7") : failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.881971 4862 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.881998 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67f7715f-3bb2-4b68-9f0c-fda6aee68678-serving-cert podName:67f7715f-3bb2-4b68-9f0c-fda6aee68678 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.381991191 +0000 UTC m=+137.914505112 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/67f7715f-3bb2-4b68-9f0c-fda6aee68678-serving-cert") pod "service-ca-operator-777779d784-s86df" (UID: "67f7715f-3bb2-4b68-9f0c-fda6aee68678") : failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.882021 4862 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.882082 4862 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.882137 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f70be661-2a14-4ba9-a271-ce98998cf14c-package-server-manager-serving-cert podName:f70be661-2a14-4ba9-a271-ce98998cf14c nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.382126855 +0000 UTC m=+137.914640776 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/f70be661-2a14-4ba9-a271-ce98998cf14c-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-wndxn" (UID: "f70be661-2a14-4ba9-a271-ce98998cf14c") : failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.882165 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-srv-cert podName:a4038318-be77-4a50-a6f2-611308bda6e7 nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.382159697 +0000 UTC m=+137.914673608 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-srv-cert") pod "olm-operator-6b444d44fb-7jpvc" (UID: "a4038318-be77-4a50-a6f2-611308bda6e7") : failed to sync secret cache: timed out waiting for the condition Nov 28 11:24:19 crc kubenswrapper[4862]: W1128 11:24:19.883950 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda2a67b7_90a1_4083_b303_9c5e7a574d58.slice/crio-8e405c3fd99a1cf49b2eded052bf6661e0d66e1b1194de1b14e615d94ab44c1c WatchSource:0}: Error finding container 8e405c3fd99a1cf49b2eded052bf6661e0d66e1b1194de1b14e615d94ab44c1c: Status 404 returned error can't find the container with id 8e405c3fd99a1cf49b2eded052bf6661e0d66e1b1194de1b14e615d94ab44c1c Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.903528 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:19 crc kubenswrapper[4862]: E1128 11:24:19.904305 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.4042836 +0000 UTC m=+137.936797521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.906353 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.913640 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.934595 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.937026 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9gvl\" (UniqueName: \"kubernetes.io/projected/1b858357-d43f-4e86-b212-479683104d02-kube-api-access-t9gvl\") pod \"openshift-config-operator-7777fb866f-kt5hh\" (UID: \"1b858357-d43f-4e86-b212-479683104d02\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.959767 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.978257 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.983913 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wt49p"] Nov 28 11:24:19 crc kubenswrapper[4862]: I1128 11:24:19.991733 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk"] Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.003694 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.007620 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.007958 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.507948007 +0000 UTC m=+138.040461928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.013439 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: W1128 11:24:20.023057 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddda64eb0_d331_41a5_a33e_323e2438c06f.slice/crio-8b61538d2b16d39c64c5e7437952dd0ff0691f943576395b4f3a0cebc007781a WatchSource:0}: Error finding container 8b61538d2b16d39c64c5e7437952dd0ff0691f943576395b4f3a0cebc007781a: Status 404 returned error can't find the container with id 8b61538d2b16d39c64c5e7437952dd0ff0691f943576395b4f3a0cebc007781a Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.038081 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.054154 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.083515 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.085030 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.097083 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.108286 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.108544 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.608490075 +0000 UTC m=+138.141004006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.113993 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.135252 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.154079 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.174650 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.201988 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.212873 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.213433 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.713417792 +0000 UTC m=+138.245931713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.213923 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.219801 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc"] Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.234417 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.254280 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.275031 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.276289 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8wmwz"] Nov 28 11:24:20 crc kubenswrapper[4862]: W1128 11:24:20.287449 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37b54aa2_ee51_4110_b25b_0739965fe748.slice/crio-62cfe4b3f8f6fc224aabe523e3a4d57851a98fdf7e2e735c1a138a1a446a8239 WatchSource:0}: Error finding container 62cfe4b3f8f6fc224aabe523e3a4d57851a98fdf7e2e735c1a138a1a446a8239: Status 404 returned error can't find the container with id 62cfe4b3f8f6fc224aabe523e3a4d57851a98fdf7e2e735c1a138a1a446a8239 Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.294995 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.310470 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh"] Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.313874 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.315624 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.815603022 +0000 UTC m=+138.348116943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.333969 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.353784 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.374204 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.394028 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.408838 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qlr6s"] Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.409841 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhl4q"] Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.414211 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.416778 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-srv-cert\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.416813 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67f7715f-3bb2-4b68-9f0c-fda6aee68678-serving-cert\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.416886 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.416918 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a1a2178-787b-4af4-a15f-225ec25b4216-config-volume\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.416972 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.417007 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67f7715f-3bb2-4b68-9f0c-fda6aee68678-config\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.417040 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a1a2178-787b-4af4-a15f-225ec25b4216-secret-volume\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.417117 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.417149 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f70be661-2a14-4ba9-a271-ce98998cf14c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wndxn\" (UID: \"f70be661-2a14-4ba9-a271-ce98998cf14c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.417181 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.417501 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:20.917486312 +0000 UTC m=+138.450000223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.418134 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67f7715f-3bb2-4b68-9f0c-fda6aee68678-config\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.418507 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a1a2178-787b-4af4-a15f-225ec25b4216-config-volume\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.418804 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.423826 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.423861 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f70be661-2a14-4ba9-a271-ce98998cf14c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wndxn\" (UID: \"f70be661-2a14-4ba9-a271-ce98998cf14c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.423880 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-srv-cert\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.424401 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a4038318-be77-4a50-a6f2-611308bda6e7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.425149 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67f7715f-3bb2-4b68-9f0c-fda6aee68678-serving-cert\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.426318 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a1a2178-787b-4af4-a15f-225ec25b4216-secret-volume\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.434561 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.453215 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.474301 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.495690 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.514017 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.518664 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.519997 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.019851458 +0000 UTC m=+138.552365389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.534467 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.553815 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.574472 4862 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.594152 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.614258 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.621616 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.622005 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.121989306 +0000 UTC m=+138.654503237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.634941 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.655054 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.660587 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" event={"ID":"da2a67b7-90a1-4083-b303-9c5e7a574d58","Type":"ContainerStarted","Data":"90ff1d40c7a6da91709846c95d15feeb668954d0296e95ca5df509b5759425e5"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.660655 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" event={"ID":"da2a67b7-90a1-4083-b303-9c5e7a574d58","Type":"ContainerStarted","Data":"8e405c3fd99a1cf49b2eded052bf6661e0d66e1b1194de1b14e615d94ab44c1c"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.664494 4862 generic.go:334] "Generic (PLEG): container finished" podID="1b858357-d43f-4e86-b212-479683104d02" containerID="1684334ce36c86058c76ae614ec662c2316c88191b72969ae8d8696bc43d1c36" exitCode=0 Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.664597 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" event={"ID":"1b858357-d43f-4e86-b212-479683104d02","Type":"ContainerDied","Data":"1684334ce36c86058c76ae614ec662c2316c88191b72969ae8d8696bc43d1c36"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.664665 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" event={"ID":"1b858357-d43f-4e86-b212-479683104d02","Type":"ContainerStarted","Data":"41cf005cb9749fc314e06598bfa10fafd606486921421cee4a3bc82daa86c293"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.666380 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" event={"ID":"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560","Type":"ContainerStarted","Data":"3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.666416 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" event={"ID":"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560","Type":"ContainerStarted","Data":"834909e4c3997062db18068324837fc15707986f43331043cfc6cb939bdfbc66"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.666709 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.668793 4862 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-bhl4q container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.668843 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" podUID="9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.669049 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" event={"ID":"3ad0f875-1ae0-4623-807e-8e88031bbd68","Type":"ContainerStarted","Data":"f09e5feeb4446820346c0c8b06eb9916818b6027fc0bd707f559d73d4b606f46"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.669087 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" event={"ID":"3ad0f875-1ae0-4623-807e-8e88031bbd68","Type":"ContainerStarted","Data":"3d01862ff6b713326d476b35aefeed7bea802af7ae8b41c7455b340fbe58dd0f"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.671893 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" event={"ID":"d7212be9-9c79-49f2-8f00-db24967afe22","Type":"ContainerStarted","Data":"28cd714c23517e2a9db0c4a34217090622a6c839fcafe7236ca4d723623fcf4c"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.671928 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" event={"ID":"d7212be9-9c79-49f2-8f00-db24967afe22","Type":"ContainerStarted","Data":"5118a1d553d96a17a26658f54214fb46707d287d45e4516888d582cb0b0e6ec8"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.672704 4862 request.go:700] Waited for 1.910207661s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/registry/token Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.674562 4862 generic.go:334] "Generic (PLEG): container finished" podID="f112a896-221b-4cfc-aef3-417a60c5be62" containerID="7b96e03da26c7aaf4d60580958001fd6422790431f1fe13d71b3197add2cf2b7" exitCode=0 Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.674630 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" event={"ID":"f112a896-221b-4cfc-aef3-417a60c5be62","Type":"ContainerDied","Data":"7b96e03da26c7aaf4d60580958001fd6422790431f1fe13d71b3197add2cf2b7"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.674651 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" event={"ID":"f112a896-221b-4cfc-aef3-417a60c5be62","Type":"ContainerStarted","Data":"0044a10fa56f2699f903a6be33121ed0caa50bdde3128d577cc277657720baac"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.677459 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8wmwz" event={"ID":"37b54aa2-ee51-4110-b25b-0739965fe748","Type":"ContainerStarted","Data":"01a9ca9cb2241b80238bc9cf325ba392d6ca2fe0a4089ce9ec84e691d6161589"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.677505 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8wmwz" event={"ID":"37b54aa2-ee51-4110-b25b-0739965fe748","Type":"ContainerStarted","Data":"62cfe4b3f8f6fc224aabe523e3a4d57851a98fdf7e2e735c1a138a1a446a8239"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.677754 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.680293 4862 patch_prober.go:28] interesting pod/console-operator-58897d9998-8wmwz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.680345 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8wmwz" podUID="37b54aa2-ee51-4110-b25b-0739965fe748" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.683445 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk" event={"ID":"dda64eb0-d331-41a5-a33e-323e2438c06f","Type":"ContainerStarted","Data":"b4b5721fda7cbee172f4221b2dbd0f9ec685b9a31453a46fae80aba307eab2e9"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.683489 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk" event={"ID":"dda64eb0-d331-41a5-a33e-323e2438c06f","Type":"ContainerStarted","Data":"b5585a72663458a35faa7acf6925fa5e6cdab6a9c2c3d08062bca6fc82ed0c3a"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.683503 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk" event={"ID":"dda64eb0-d331-41a5-a33e-323e2438c06f","Type":"ContainerStarted","Data":"8b61538d2b16d39c64c5e7437952dd0ff0691f943576395b4f3a0cebc007781a"} Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.702763 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7plxg\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-kube-api-access-7plxg\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.707081 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wcjf\" (UniqueName: \"kubernetes.io/projected/d275feb5-1839-4942-9330-c4f5c47cc0fe-kube-api-access-2wcjf\") pod \"etcd-operator-b45778765-5hjfs\" (UID: \"d275feb5-1839-4942-9330-c4f5c47cc0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.713485 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.723456 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.723858 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.223843905 +0000 UTC m=+138.756357826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.735197 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.754680 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.774738 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.809059 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.812518 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-bound-sa-token\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.824939 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.825376 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.325359674 +0000 UTC m=+138.857873595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.846725 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8fwm\" (UniqueName: \"kubernetes.io/projected/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-kube-api-access-k8fwm\") pod \"console-f9d7485db-pzshw\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.850865 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djt7v\" (UniqueName: \"kubernetes.io/projected/56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f-kube-api-access-djt7v\") pod \"kube-storage-version-migrator-operator-b67b599dd-tx94r\" (UID: \"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.875646 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6lq7\" (UniqueName: \"kubernetes.io/projected/90054579-c9cd-44f7-8d0d-bca5c4bd6e20-kube-api-access-l6lq7\") pod \"apiserver-7bbb656c7d-cld25\" (UID: \"90054579-c9cd-44f7-8d0d-bca5c4bd6e20\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.888990 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4wk4\" (UniqueName: \"kubernetes.io/projected/c34ee8a3-cf6f-4d17-9638-8950c56b9fc8-kube-api-access-q4wk4\") pod \"multus-admission-controller-857f4d67dd-qlsvq\" (UID: \"c34ee8a3-cf6f-4d17-9638-8950c56b9fc8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.895314 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.917292 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfklc\" (UniqueName: \"kubernetes.io/projected/3b577ecb-1896-4699-bd7e-33929f2cf9a3-kube-api-access-cfklc\") pod \"marketplace-operator-79b997595-nbzb4\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.925445 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.925712 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.425686045 +0000 UTC m=+138.958199966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.925889 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:20 crc kubenswrapper[4862]: E1128 11:24:20.926228 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.426220992 +0000 UTC m=+138.958734913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.940002 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x72m\" (UniqueName: \"kubernetes.io/projected/9b3ee9cc-2155-4e40-a8e3-6ed95db6962b-kube-api-access-4x72m\") pod \"cluster-samples-operator-665b6dd947-9f6sc\" (UID: \"9b3ee9cc-2155-4e40-a8e3-6ed95db6962b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.948836 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.952807 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mth8\" (UniqueName: \"kubernetes.io/projected/5c0e4c02-c077-4c5e-bc39-0601250102b3-kube-api-access-7mth8\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.972142 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll696\" (UniqueName: \"kubernetes.io/projected/3a1a2178-787b-4af4-a15f-225ec25b4216-kube-api-access-ll696\") pod \"collect-profiles-29405475-qkjm5\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.993400 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:20 crc kubenswrapper[4862]: I1128 11:24:20.994417 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f45f\" (UniqueName: \"kubernetes.io/projected/a4038318-be77-4a50-a6f2-611308bda6e7-kube-api-access-6f45f\") pod \"olm-operator-6b444d44fb-7jpvc\" (UID: \"a4038318-be77-4a50-a6f2-611308bda6e7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.006583 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.009234 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce5a45b8-754f-46aa-a0eb-8325aa5a3868-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gvpct\" (UID: \"ce5a45b8-754f-46aa-a0eb-8325aa5a3868\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.026437 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.026835 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.526821581 +0000 UTC m=+139.059335502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.030787 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5c0e4c02-c077-4c5e-bc39-0601250102b3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bp9bf\" (UID: \"5c0e4c02-c077-4c5e-bc39-0601250102b3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.039319 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.064238 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54gdk\" (UniqueName: \"kubernetes.io/projected/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-kube-api-access-54gdk\") pod \"oauth-openshift-558db77b4-g5fhw\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.090400 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qxfz\" (UniqueName: \"kubernetes.io/projected/170cfb20-5cf0-4050-944b-6b17b0e2decc-kube-api-access-5qxfz\") pod \"dns-operator-744455d44c-7mhnb\" (UID: \"170cfb20-5cf0-4050-944b-6b17b0e2decc\") " pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.091643 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5hjfs"] Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.103368 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljvzf\" (UniqueName: \"kubernetes.io/projected/f70be661-2a14-4ba9-a271-ce98998cf14c-kube-api-access-ljvzf\") pod \"package-server-manager-789f6589d5-wndxn\" (UID: \"f70be661-2a14-4ba9-a271-ce98998cf14c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.120038 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.123441 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.124371 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psxr8\" (UniqueName: \"kubernetes.io/projected/faca5cde-65a6-429e-816c-a97fa57c5bf1-kube-api-access-psxr8\") pod \"downloads-7954f5f757-45tfc\" (UID: \"faca5cde-65a6-429e-816c-a97fa57c5bf1\") " pod="openshift-console/downloads-7954f5f757-45tfc" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.128406 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.128767 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.628752093 +0000 UTC m=+139.161266014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.132616 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-45tfc" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.139296 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.143239 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sscpn\" (UniqueName: \"kubernetes.io/projected/ee1f3625-a084-4e0d-99af-627305b0f8f2-kube-api-access-sscpn\") pod \"route-controller-manager-6576b87f9c-x5qls\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.157829 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.164406 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.164950 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-288ph\" (UniqueName: \"kubernetes.io/projected/d14c0311-e7c1-4464-8872-f5e28b3ee9a2-kube-api-access-288ph\") pod \"machine-approver-56656f9798-rgpc4\" (UID: \"d14c0311-e7c1-4464-8872-f5e28b3ee9a2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.207180 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxxpb\" (UniqueName: \"kubernetes.io/projected/67f7715f-3bb2-4b68-9f0c-fda6aee68678-kube-api-access-fxxpb\") pod \"service-ca-operator-777779d784-s86df\" (UID: \"67f7715f-3bb2-4b68-9f0c-fda6aee68678\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.208950 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.226270 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.226702 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.228261 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r"] Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.230971 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.231123 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.731103548 +0000 UTC m=+139.263617469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.233647 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.233906 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.733897767 +0000 UTC m=+139.266411688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.234137 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.235533 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f40b08d-cf89-4ff6-a766-a38aab93f8dd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xw6qk\" (UID: \"3f40b08d-cf89-4ff6-a766-a38aab93f8dd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.237758 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.252162 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.253402 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.276218 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.295345 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.299025 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.314627 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.315043 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.335227 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.335687 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.835663884 +0000 UTC m=+139.368177805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: W1128 11:24:21.367199 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56a5ba03_2ecb_48b7_bff6_5fa97bd4e07f.slice/crio-fe37ff43875b0ab72e825c6759b36b849adc811a4eacd5b534d29a2f3c20fd7e WatchSource:0}: Error finding container fe37ff43875b0ab72e825c6759b36b849adc811a4eacd5b534d29a2f3c20fd7e: Status 404 returned error can't find the container with id fe37ff43875b0ab72e825c6759b36b849adc811a4eacd5b534d29a2f3c20fd7e Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.436900 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-stats-auth\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.437788 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a23fb95f-e0c8-4681-b797-035639fed352-profile-collector-cert\") pod \"catalog-operator-68c6474976-w9lkd\" (UID: \"a23fb95f-e0c8-4681-b797-035639fed352\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.438209 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.438236 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f4da0ab8-bdd5-44eb-ac21-626d7c983fa7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-s926b\" (UID: \"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.438556 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf5qd\" (UniqueName: \"kubernetes.io/projected/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-kube-api-access-qf5qd\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.438613 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnxt7\" (UniqueName: \"kubernetes.io/projected/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-kube-api-access-xnxt7\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.438636 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-service-ca-bundle\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.438691 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g48qq\" (UniqueName: \"kubernetes.io/projected/a23fb95f-e0c8-4681-b797-035639fed352-kube-api-access-g48qq\") pod \"catalog-operator-68c6474976-w9lkd\" (UID: \"a23fb95f-e0c8-4681-b797-035639fed352\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.438710 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d69v\" (UniqueName: \"kubernetes.io/projected/8fd16cd6-9d43-4843-b091-aaf0686c9b94-kube-api-access-5d69v\") pod \"openshift-controller-manager-operator-756b6f6bc6-m6kbw\" (UID: \"8fd16cd6-9d43-4843-b091-aaf0686c9b94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.440953 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.441023 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.441532 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:21.941518521 +0000 UTC m=+139.474032442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.442427 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw692\" (UniqueName: \"kubernetes.io/projected/f4da0ab8-bdd5-44eb-ac21-626d7c983fa7-kube-api-access-jw692\") pod \"machine-config-controller-84d6567774-s926b\" (UID: \"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.442975 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd16cd6-9d43-4843-b091-aaf0686c9b94-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-m6kbw\" (UID: \"8fd16cd6-9d43-4843-b091-aaf0686c9b94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.443030 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd16cd6-9d43-4843-b091-aaf0686c9b94-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-m6kbw\" (UID: \"8fd16cd6-9d43-4843-b091-aaf0686c9b94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.443199 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-default-certificate\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.443460 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brbqt\" (UniqueName: \"kubernetes.io/projected/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-kube-api-access-brbqt\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.443974 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-metrics-certs\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.444044 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.444084 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-webhook-cert\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.444380 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f4da0ab8-bdd5-44eb-ac21-626d7c983fa7-proxy-tls\") pod \"machine-config-controller-84d6567774-s926b\" (UID: \"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.444505 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9n8z\" (UniqueName: \"kubernetes.io/projected/47f57404-6c91-49f7-bfc5-8a4c2ddaca35-kube-api-access-c9n8z\") pod \"control-plane-machine-set-operator-78cbb6b69f-9xwmh\" (UID: \"47f57404-6c91-49f7-bfc5-8a4c2ddaca35\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.444820 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a23fb95f-e0c8-4681-b797-035639fed352-srv-cert\") pod \"catalog-operator-68c6474976-w9lkd\" (UID: \"a23fb95f-e0c8-4681-b797-035639fed352\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.446441 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/47f57404-6c91-49f7-bfc5-8a4c2ddaca35-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9xwmh\" (UID: \"47f57404-6c91-49f7-bfc5-8a4c2ddaca35\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.446666 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-apiservice-cert\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.446875 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-tmpfs\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.447231 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.540934 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5"] Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551338 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551488 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f4da0ab8-bdd5-44eb-ac21-626d7c983fa7-proxy-tls\") pod \"machine-config-controller-84d6567774-s926b\" (UID: \"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551523 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-webhook-cert\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551549 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/daf7469b-a0bd-4682-a6df-502827ae1306-proxy-tls\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551577 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/daf7469b-a0bd-4682-a6df-502827ae1306-images\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551600 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bc932a43-68f4-4582-8894-6fcae516668c-cert\") pod \"ingress-canary-d5qm9\" (UID: \"bc932a43-68f4-4582-8894-6fcae516668c\") " pod="openshift-ingress-canary/ingress-canary-d5qm9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551624 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80fef19a-293d-41fa-9ece-82f411f960a0-config-volume\") pod \"dns-default-kpm6d\" (UID: \"80fef19a-293d-41fa-9ece-82f411f960a0\") " pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551644 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80fef19a-293d-41fa-9ece-82f411f960a0-metrics-tls\") pod \"dns-default-kpm6d\" (UID: \"80fef19a-293d-41fa-9ece-82f411f960a0\") " pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551691 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9n8z\" (UniqueName: \"kubernetes.io/projected/47f57404-6c91-49f7-bfc5-8a4c2ddaca35-kube-api-access-c9n8z\") pod \"control-plane-machine-set-operator-78cbb6b69f-9xwmh\" (UID: \"47f57404-6c91-49f7-bfc5-8a4c2ddaca35\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551746 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a23fb95f-e0c8-4681-b797-035639fed352-srv-cert\") pod \"catalog-operator-68c6474976-w9lkd\" (UID: \"a23fb95f-e0c8-4681-b797-035639fed352\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551811 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f62f275b-b840-4b88-8992-a2f6ecc08ab0-signing-cabundle\") pod \"service-ca-9c57cc56f-q7bbb\" (UID: \"f62f275b-b840-4b88-8992-a2f6ecc08ab0\") " pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551833 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf6tx\" (UniqueName: \"kubernetes.io/projected/d15923b5-8e5e-4c97-9137-2e84e65d63c4-kube-api-access-cf6tx\") pod \"machine-config-server-nqpc6\" (UID: \"d15923b5-8e5e-4c97-9137-2e84e65d63c4\") " pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551907 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/47f57404-6c91-49f7-bfc5-8a4c2ddaca35-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9xwmh\" (UID: \"47f57404-6c91-49f7-bfc5-8a4c2ddaca35\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551948 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-apiservice-cert\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.551968 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-tmpfs\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552001 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln7w6\" (UniqueName: \"kubernetes.io/projected/daf7469b-a0bd-4682-a6df-502827ae1306-kube-api-access-ln7w6\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552022 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d15923b5-8e5e-4c97-9137-2e84e65d63c4-node-bootstrap-token\") pod \"machine-config-server-nqpc6\" (UID: \"d15923b5-8e5e-4c97-9137-2e84e65d63c4\") " pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552080 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d15923b5-8e5e-4c97-9137-2e84e65d63c4-certs\") pod \"machine-config-server-nqpc6\" (UID: \"d15923b5-8e5e-4c97-9137-2e84e65d63c4\") " pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552137 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-stats-auth\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552222 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-registration-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552265 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a23fb95f-e0c8-4681-b797-035639fed352-profile-collector-cert\") pod \"catalog-operator-68c6474976-w9lkd\" (UID: \"a23fb95f-e0c8-4681-b797-035639fed352\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552283 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-mountpoint-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552311 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-csi-data-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552331 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-plugins-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552381 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6mlp\" (UniqueName: \"kubernetes.io/projected/bc932a43-68f4-4582-8894-6fcae516668c-kube-api-access-q6mlp\") pod \"ingress-canary-d5qm9\" (UID: \"bc932a43-68f4-4582-8894-6fcae516668c\") " pod="openshift-ingress-canary/ingress-canary-d5qm9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552413 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/daf7469b-a0bd-4682-a6df-502827ae1306-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552471 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552501 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f4da0ab8-bdd5-44eb-ac21-626d7c983fa7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-s926b\" (UID: \"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552520 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf5qd\" (UniqueName: \"kubernetes.io/projected/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-kube-api-access-qf5qd\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552816 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-service-ca-bundle\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552877 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnxt7\" (UniqueName: \"kubernetes.io/projected/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-kube-api-access-xnxt7\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552933 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g48qq\" (UniqueName: \"kubernetes.io/projected/a23fb95f-e0c8-4681-b797-035639fed352-kube-api-access-g48qq\") pod \"catalog-operator-68c6474976-w9lkd\" (UID: \"a23fb95f-e0c8-4681-b797-035639fed352\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.552958 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d69v\" (UniqueName: \"kubernetes.io/projected/8fd16cd6-9d43-4843-b091-aaf0686c9b94-kube-api-access-5d69v\") pod \"openshift-controller-manager-operator-756b6f6bc6-m6kbw\" (UID: \"8fd16cd6-9d43-4843-b091-aaf0686c9b94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553040 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553201 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw692\" (UniqueName: \"kubernetes.io/projected/f4da0ab8-bdd5-44eb-ac21-626d7c983fa7-kube-api-access-jw692\") pod \"machine-config-controller-84d6567774-s926b\" (UID: \"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553296 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd16cd6-9d43-4843-b091-aaf0686c9b94-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-m6kbw\" (UID: \"8fd16cd6-9d43-4843-b091-aaf0686c9b94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553336 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4qdx\" (UniqueName: \"kubernetes.io/projected/f62f275b-b840-4b88-8992-a2f6ecc08ab0-kube-api-access-s4qdx\") pod \"service-ca-9c57cc56f-q7bbb\" (UID: \"f62f275b-b840-4b88-8992-a2f6ecc08ab0\") " pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553368 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd16cd6-9d43-4843-b091-aaf0686c9b94-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-m6kbw\" (UID: \"8fd16cd6-9d43-4843-b091-aaf0686c9b94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553386 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-socket-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553419 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xvlv\" (UniqueName: \"kubernetes.io/projected/80fef19a-293d-41fa-9ece-82f411f960a0-kube-api-access-7xvlv\") pod \"dns-default-kpm6d\" (UID: \"80fef19a-293d-41fa-9ece-82f411f960a0\") " pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553463 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-default-certificate\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553484 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xlxj\" (UniqueName: \"kubernetes.io/projected/b2656085-4bdf-495b-a4e7-92042752c35d-kube-api-access-5xlxj\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553538 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brbqt\" (UniqueName: \"kubernetes.io/projected/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-kube-api-access-brbqt\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553558 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f62f275b-b840-4b88-8992-a2f6ecc08ab0-signing-key\") pod \"service-ca-9c57cc56f-q7bbb\" (UID: \"f62f275b-b840-4b88-8992-a2f6ecc08ab0\") " pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553658 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-metrics-certs\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.553736 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.554214 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.054194604 +0000 UTC m=+139.586708525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.563339 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f4da0ab8-bdd5-44eb-ac21-626d7c983fa7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-s926b\" (UID: \"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.565611 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-service-ca-bundle\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.582868 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd16cd6-9d43-4843-b091-aaf0686c9b94-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-m6kbw\" (UID: \"8fd16cd6-9d43-4843-b091-aaf0686c9b94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.585364 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-webhook-cert\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.591651 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-stats-auth\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.636442 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-apiservice-cert\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.636874 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-default-certificate\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.638612 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f4da0ab8-bdd5-44eb-ac21-626d7c983fa7-proxy-tls\") pod \"machine-config-controller-84d6567774-s926b\" (UID: \"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.640779 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf5qd\" (UniqueName: \"kubernetes.io/projected/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-kube-api-access-qf5qd\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.643203 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a23fb95f-e0c8-4681-b797-035639fed352-srv-cert\") pod \"catalog-operator-68c6474976-w9lkd\" (UID: \"a23fb95f-e0c8-4681-b797-035639fed352\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.645068 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/269efc51-3dfb-4888-bc8f-4f681ac5d2c6-metrics-certs\") pod \"router-default-5444994796-2bhn7\" (UID: \"269efc51-3dfb-4888-bc8f-4f681ac5d2c6\") " pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.645794 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/47f57404-6c91-49f7-bfc5-8a4c2ddaca35-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9xwmh\" (UID: \"47f57404-6c91-49f7-bfc5-8a4c2ddaca35\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.646893 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a23fb95f-e0c8-4681-b797-035639fed352-profile-collector-cert\") pod \"catalog-operator-68c6474976-w9lkd\" (UID: \"a23fb95f-e0c8-4681-b797-035639fed352\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.650467 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nbzb4"] Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.650706 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pzshw"] Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.654948 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf6tx\" (UniqueName: \"kubernetes.io/projected/d15923b5-8e5e-4c97-9137-2e84e65d63c4-kube-api-access-cf6tx\") pod \"machine-config-server-nqpc6\" (UID: \"d15923b5-8e5e-4c97-9137-2e84e65d63c4\") " pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655006 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln7w6\" (UniqueName: \"kubernetes.io/projected/daf7469b-a0bd-4682-a6df-502827ae1306-kube-api-access-ln7w6\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655031 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d15923b5-8e5e-4c97-9137-2e84e65d63c4-node-bootstrap-token\") pod \"machine-config-server-nqpc6\" (UID: \"d15923b5-8e5e-4c97-9137-2e84e65d63c4\") " pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655069 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d15923b5-8e5e-4c97-9137-2e84e65d63c4-certs\") pod \"machine-config-server-nqpc6\" (UID: \"d15923b5-8e5e-4c97-9137-2e84e65d63c4\") " pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655117 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-registration-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655139 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-mountpoint-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655161 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-csi-data-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655181 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-plugins-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655209 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6mlp\" (UniqueName: \"kubernetes.io/projected/bc932a43-68f4-4582-8894-6fcae516668c-kube-api-access-q6mlp\") pod \"ingress-canary-d5qm9\" (UID: \"bc932a43-68f4-4582-8894-6fcae516668c\") " pod="openshift-ingress-canary/ingress-canary-d5qm9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655230 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/daf7469b-a0bd-4682-a6df-502827ae1306-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655309 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655357 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4qdx\" (UniqueName: \"kubernetes.io/projected/f62f275b-b840-4b88-8992-a2f6ecc08ab0-kube-api-access-s4qdx\") pod \"service-ca-9c57cc56f-q7bbb\" (UID: \"f62f275b-b840-4b88-8992-a2f6ecc08ab0\") " pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655374 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-socket-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655394 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xvlv\" (UniqueName: \"kubernetes.io/projected/80fef19a-293d-41fa-9ece-82f411f960a0-kube-api-access-7xvlv\") pod \"dns-default-kpm6d\" (UID: \"80fef19a-293d-41fa-9ece-82f411f960a0\") " pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655415 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xlxj\" (UniqueName: \"kubernetes.io/projected/b2656085-4bdf-495b-a4e7-92042752c35d-kube-api-access-5xlxj\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655443 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f62f275b-b840-4b88-8992-a2f6ecc08ab0-signing-key\") pod \"service-ca-9c57cc56f-q7bbb\" (UID: \"f62f275b-b840-4b88-8992-a2f6ecc08ab0\") " pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655487 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/daf7469b-a0bd-4682-a6df-502827ae1306-proxy-tls\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655506 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/daf7469b-a0bd-4682-a6df-502827ae1306-images\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655527 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bc932a43-68f4-4582-8894-6fcae516668c-cert\") pod \"ingress-canary-d5qm9\" (UID: \"bc932a43-68f4-4582-8894-6fcae516668c\") " pod="openshift-ingress-canary/ingress-canary-d5qm9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655545 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80fef19a-293d-41fa-9ece-82f411f960a0-config-volume\") pod \"dns-default-kpm6d\" (UID: \"80fef19a-293d-41fa-9ece-82f411f960a0\") " pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655562 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80fef19a-293d-41fa-9ece-82f411f960a0-metrics-tls\") pod \"dns-default-kpm6d\" (UID: \"80fef19a-293d-41fa-9ece-82f411f960a0\") " pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655564 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-csi-data-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.655602 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f62f275b-b840-4b88-8992-a2f6ecc08ab0-signing-cabundle\") pod \"service-ca-9c57cc56f-q7bbb\" (UID: \"f62f275b-b840-4b88-8992-a2f6ecc08ab0\") " pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.656310 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f62f275b-b840-4b88-8992-a2f6ecc08ab0-signing-cabundle\") pod \"service-ca-9c57cc56f-q7bbb\" (UID: \"f62f275b-b840-4b88-8992-a2f6ecc08ab0\") " pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.656751 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-plugins-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.659831 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/daf7469b-a0bd-4682-a6df-502827ae1306-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.660335 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.160317989 +0000 UTC m=+139.692831910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.660918 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-socket-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.661073 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d15923b5-8e5e-4c97-9137-2e84e65d63c4-node-bootstrap-token\") pod \"machine-config-server-nqpc6\" (UID: \"d15923b5-8e5e-4c97-9137-2e84e65d63c4\") " pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.661195 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-registration-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.661234 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b2656085-4bdf-495b-a4e7-92042752c35d-mountpoint-dir\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.662463 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9n8z\" (UniqueName: \"kubernetes.io/projected/47f57404-6c91-49f7-bfc5-8a4c2ddaca35-kube-api-access-c9n8z\") pod \"control-plane-machine-set-operator-78cbb6b69f-9xwmh\" (UID: \"47f57404-6c91-49f7-bfc5-8a4c2ddaca35\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.663249 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80fef19a-293d-41fa-9ece-82f411f960a0-config-volume\") pod \"dns-default-kpm6d\" (UID: \"80fef19a-293d-41fa-9ece-82f411f960a0\") " pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.664295 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/daf7469b-a0bd-4682-a6df-502827ae1306-images\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.671574 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f62f275b-b840-4b88-8992-a2f6ecc08ab0-signing-key\") pod \"service-ca-9c57cc56f-q7bbb\" (UID: \"f62f275b-b840-4b88-8992-a2f6ecc08ab0\") " pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.688522 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bc932a43-68f4-4582-8894-6fcae516668c-cert\") pod \"ingress-canary-d5qm9\" (UID: \"bc932a43-68f4-4582-8894-6fcae516668c\") " pod="openshift-ingress-canary/ingress-canary-d5qm9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.699926 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80fef19a-293d-41fa-9ece-82f411f960a0-metrics-tls\") pod \"dns-default-kpm6d\" (UID: \"80fef19a-293d-41fa-9ece-82f411f960a0\") " pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.700660 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-tmpfs\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.701320 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d15923b5-8e5e-4c97-9137-2e84e65d63c4-certs\") pod \"machine-config-server-nqpc6\" (UID: \"d15923b5-8e5e-4c97-9137-2e84e65d63c4\") " pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.702615 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g48qq\" (UniqueName: \"kubernetes.io/projected/a23fb95f-e0c8-4681-b797-035639fed352-kube-api-access-g48qq\") pod \"catalog-operator-68c6474976-w9lkd\" (UID: \"a23fb95f-e0c8-4681-b797-035639fed352\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.702974 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd16cd6-9d43-4843-b091-aaf0686c9b94-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-m6kbw\" (UID: \"8fd16cd6-9d43-4843-b091-aaf0686c9b94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.706741 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brbqt\" (UniqueName: \"kubernetes.io/projected/cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed-kube-api-access-brbqt\") pod \"packageserver-d55dfcdfc-kl4jk\" (UID: \"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.711786 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/daf7469b-a0bd-4682-a6df-502827ae1306-proxy-tls\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.713104 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d69v\" (UniqueName: \"kubernetes.io/projected/8fd16cd6-9d43-4843-b091-aaf0686c9b94-kube-api-access-5d69v\") pod \"openshift-controller-manager-operator-756b6f6bc6-m6kbw\" (UID: \"8fd16cd6-9d43-4843-b091-aaf0686c9b94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.721553 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc"] Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.727471 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw692\" (UniqueName: \"kubernetes.io/projected/f4da0ab8-bdd5-44eb-ac21-626d7c983fa7-kube-api-access-jw692\") pod \"machine-config-controller-84d6567774-s926b\" (UID: \"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.728531 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.731969 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf6tx\" (UniqueName: \"kubernetes.io/projected/d15923b5-8e5e-4c97-9137-2e84e65d63c4-kube-api-access-cf6tx\") pod \"machine-config-server-nqpc6\" (UID: \"d15923b5-8e5e-4c97-9137-2e84e65d63c4\") " pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.739796 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.743813 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnxt7\" (UniqueName: \"kubernetes.io/projected/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-kube-api-access-xnxt7\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.759497 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.760788 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.260759504 +0000 UTC m=+139.793273425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.762665 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.770577 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.270563845 +0000 UTC m=+139.803077766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.772364 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/33f176fd-b230-4a6c-993f-68ff9a8e3cd7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5gcl2\" (UID: \"33f176fd-b230-4a6c-993f-68ff9a8e3cd7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.774561 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln7w6\" (UniqueName: \"kubernetes.io/projected/daf7469b-a0bd-4682-a6df-502827ae1306-kube-api-access-ln7w6\") pod \"machine-config-operator-74547568cd-8hqk9\" (UID: \"daf7469b-a0bd-4682-a6df-502827ae1306\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.775451 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" event={"ID":"d275feb5-1839-4942-9330-c4f5c47cc0fe","Type":"ContainerStarted","Data":"ee9c4b2a5636b2f377ac1ca6a6cd72a3475ef2972b04e0b4a571332c4bb5b890"} Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.776668 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25"] Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.777682 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" event={"ID":"1b858357-d43f-4e86-b212-479683104d02","Type":"ContainerStarted","Data":"fe4d01f73e124f797aa9b9a9f6f38b9e338a01e0fe78478f9838d04f232d59d9"} Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.777815 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.781250 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" event={"ID":"3ad0f875-1ae0-4623-807e-8e88031bbd68","Type":"ContainerStarted","Data":"2673cc74d4768d998f8ebb69b1fa5e4cd7e8d4ce6d9eba87ec9134366f902de7"} Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.797600 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.800203 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.805919 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" event={"ID":"f112a896-221b-4cfc-aef3-417a60c5be62","Type":"ContainerStarted","Data":"3adca9072d4660e9d2e47eea7fdc37e1822c3d4f66272017fe18eb81c828fb40"} Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.808652 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4qdx\" (UniqueName: \"kubernetes.io/projected/f62f275b-b840-4b88-8992-a2f6ecc08ab0-kube-api-access-s4qdx\") pod \"service-ca-9c57cc56f-q7bbb\" (UID: \"f62f275b-b840-4b88-8992-a2f6ecc08ab0\") " pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.808663 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xlxj\" (UniqueName: \"kubernetes.io/projected/b2656085-4bdf-495b-a4e7-92042752c35d-kube-api-access-5xlxj\") pod \"csi-hostpathplugin-l8mhs\" (UID: \"b2656085-4bdf-495b-a4e7-92042752c35d\") " pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.815786 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" event={"ID":"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f","Type":"ContainerStarted","Data":"fe37ff43875b0ab72e825c6759b36b849adc811a4eacd5b534d29a2f3c20fd7e"} Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.822410 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.823166 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.830661 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.840956 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xvlv\" (UniqueName: \"kubernetes.io/projected/80fef19a-293d-41fa-9ece-82f411f960a0-kube-api-access-7xvlv\") pod \"dns-default-kpm6d\" (UID: \"80fef19a-293d-41fa-9ece-82f411f960a0\") " pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.857035 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-8wmwz" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.865501 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.871922 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.873376 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.373351344 +0000 UTC m=+139.905865265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.873485 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.883716 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6mlp\" (UniqueName: \"kubernetes.io/projected/bc932a43-68f4-4582-8894-6fcae516668c-kube-api-access-q6mlp\") pod \"ingress-canary-d5qm9\" (UID: \"bc932a43-68f4-4582-8894-6fcae516668c\") " pod="openshift-ingress-canary/ingress-canary-d5qm9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.928403 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.950326 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.975363 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-nqpc6" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.975702 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.981195 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:21 crc kubenswrapper[4862]: E1128 11:24:21.981618 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.481602857 +0000 UTC m=+140.014116778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:21 crc kubenswrapper[4862]: I1128 11:24:21.988231 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7mhnb"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.004197 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.039675 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.040137 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.062124 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-d5qm9" Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.080011 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g5fhw"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.093663 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:22 crc kubenswrapper[4862]: E1128 11:24:22.094126 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.594082744 +0000 UTC m=+140.126596665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.119264 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-45tfc"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.167728 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8bqmb" podStartSLOduration=121.167695816 podStartE2EDuration="2m1.167695816s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:22.133638962 +0000 UTC m=+139.666152883" watchObservedRunningTime="2025-11-28 11:24:22.167695816 +0000 UTC m=+139.700209737" Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.194731 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:22 crc kubenswrapper[4862]: E1128 11:24:22.195056 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.695045235 +0000 UTC m=+140.227559156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.200682 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mlwqk" podStartSLOduration=120.200657054 podStartE2EDuration="2m0.200657054s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:22.199841988 +0000 UTC m=+139.732355919" watchObservedRunningTime="2025-11-28 11:24:22.200657054 +0000 UTC m=+139.733170985" Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.291264 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qlsvq"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.314796 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:22 crc kubenswrapper[4862]: E1128 11:24:22.315321 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.81530659 +0000 UTC m=+140.347820501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.392163 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.416871 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:22 crc kubenswrapper[4862]: E1128 11:24:22.417222 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:22.917208781 +0000 UTC m=+140.449722702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.420720 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.515465 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.519593 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:22 crc kubenswrapper[4862]: E1128 11:24:22.519920 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.019900936 +0000 UTC m=+140.552414867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:22 crc kubenswrapper[4862]: W1128 11:24:22.616384 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce5a45b8_754f_46aa_a0eb_8325aa5a3868.slice/crio-8ae5181af159175e5c533ae82602d6c723074865266a3ea6fd9b992c960b7fbc WatchSource:0}: Error finding container 8ae5181af159175e5c533ae82602d6c723074865266a3ea6fd9b992c960b7fbc: Status 404 returned error can't find the container with id 8ae5181af159175e5c533ae82602d6c723074865266a3ea6fd9b992c960b7fbc Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.626668 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:22 crc kubenswrapper[4862]: E1128 11:24:22.627281 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.127265631 +0000 UTC m=+140.659779542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.640550 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.727430 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:22 crc kubenswrapper[4862]: E1128 11:24:22.728244 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.228226052 +0000 UTC m=+140.760739963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.769382 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-s86df"] Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.835336 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:22 crc kubenswrapper[4862]: E1128 11:24:22.835750 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.335736751 +0000 UTC m=+140.868250672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.909766 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" event={"ID":"3a1a2178-787b-4af4-a15f-225ec25b4216","Type":"ContainerStarted","Data":"ee937bcba2a2835eaa2fd43dc9bc9f65a9158b80a528d396d5291454cbefb273"} Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.910156 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" event={"ID":"ee1f3625-a084-4e0d-99af-627305b0f8f2","Type":"ContainerStarted","Data":"097b4b645db34dd3616fe180841b937e52c189be76d11fb9eada62279fdb38e9"} Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.910167 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" event={"ID":"a4038318-be77-4a50-a6f2-611308bda6e7","Type":"ContainerStarted","Data":"774c9eefa3063f96796b3648bc36bffe6c778059ee06b998eaea4917c3d61041"} Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.936143 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:22 crc kubenswrapper[4862]: E1128 11:24:22.936476 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.436460224 +0000 UTC m=+140.968974145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.940963 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" event={"ID":"f112a896-221b-4cfc-aef3-417a60c5be62","Type":"ContainerStarted","Data":"fe1caf79773d618498baa558ef280bd2bf3065e83985e139fc1f9e0ef8757807"} Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.951728 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-qlr6s" podStartSLOduration=120.951711199 podStartE2EDuration="2m0.951711199s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:22.949614152 +0000 UTC m=+140.482128074" watchObservedRunningTime="2025-11-28 11:24:22.951711199 +0000 UTC m=+140.484225120" Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.963526 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" event={"ID":"56a5ba03-2ecb-48b7-bff6-5fa97bd4e07f","Type":"ContainerStarted","Data":"e687a189c808f352dcd09ed3b694e0ec5dfea69674193d60add51eea59f17824"} Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.991525 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" event={"ID":"d14c0311-e7c1-4464-8872-f5e28b3ee9a2","Type":"ContainerStarted","Data":"b8caf85ff4fd2745aa05bd753a9f9a396b6919a82f72bbaf277c2b7c8bc7359a"} Nov 28 11:24:22 crc kubenswrapper[4862]: I1128 11:24:22.995673 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" event={"ID":"c34ee8a3-cf6f-4d17-9638-8950c56b9fc8","Type":"ContainerStarted","Data":"7a700cc00b23fb56a588c79af5136182dfc49a31ee82cba3cd599456c57f1c33"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.003410 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" event={"ID":"9b3ee9cc-2155-4e40-a8e3-6ed95db6962b","Type":"ContainerStarted","Data":"1533675718a571f0841a9942516cdc3416a2c37ea3c9ab1a304aeda0d0a28d3d"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.047601 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.048991 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.548978133 +0000 UTC m=+141.081492054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.065462 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" event={"ID":"d275feb5-1839-4942-9330-c4f5c47cc0fe","Type":"ContainerStarted","Data":"de6af24d4ced4fb6d3ab4591b3656af167bd313baf7bfbebe78e6d56c1cb4196"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.093932 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.094626 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-45tfc" event={"ID":"faca5cde-65a6-429e-816c-a97fa57c5bf1","Type":"ContainerStarted","Data":"f20b70e91870b3464beada681e9528200e06003cf56ebc08caecf0ac5e056ca2"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.132053 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pzshw" event={"ID":"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7","Type":"ContainerStarted","Data":"3db17258d2ca71f27c91c0077b3e217702126185806a5630b324097cee3146d9"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.157260 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.157343 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.657326999 +0000 UTC m=+141.189840910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.158519 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.160885 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.660873181 +0000 UTC m=+141.193387102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.167857 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" event={"ID":"5c0e4c02-c077-4c5e-bc39-0601250102b3","Type":"ContainerStarted","Data":"9f745b819a8d4f7277c140cc62b5f9c88b153932a92fbdf63339fa650ab0fb29"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.182374 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh"] Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.182496 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" event={"ID":"170cfb20-5cf0-4050-944b-6b17b0e2decc","Type":"ContainerStarted","Data":"f579c68f88b12dddec5146cfa3b95dd843476fcccebe60e2a6fe856b5a1498fb"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.230949 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kt5hh" podStartSLOduration=121.230916299 podStartE2EDuration="2m1.230916299s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:23.228727339 +0000 UTC m=+140.761241260" watchObservedRunningTime="2025-11-28 11:24:23.230916299 +0000 UTC m=+140.763430220" Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.240071 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2bhn7" event={"ID":"269efc51-3dfb-4888-bc8f-4f681ac5d2c6","Type":"ContainerStarted","Data":"b06f712877b70cba29b169a12c7032e859458f58e4d700f201a4083cdbc90f21"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.242738 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" event={"ID":"3f40b08d-cf89-4ff6-a766-a38aab93f8dd","Type":"ContainerStarted","Data":"ae16020e00c4375288793a014fa940ea474b870dd5191773f8296f11ad08272f"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.253719 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" event={"ID":"90054579-c9cd-44f7-8d0d-bca5c4bd6e20","Type":"ContainerStarted","Data":"098ee05557b0d1f72f5747a184c4a89111876c3c1f3e94d2023447fb2bb42d03"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.256361 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" event={"ID":"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af","Type":"ContainerStarted","Data":"8f9494c5ffa82a680ce68bb20b0a675b3dfb162273cee72a2df964cdae9010ae"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.259358 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.259715 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.759700044 +0000 UTC m=+141.292213965 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.260996 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" event={"ID":"ce5a45b8-754f-46aa-a0eb-8325aa5a3868","Type":"ContainerStarted","Data":"8ae5181af159175e5c533ae82602d6c723074865266a3ea6fd9b992c960b7fbc"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.267138 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" event={"ID":"3b577ecb-1896-4699-bd7e-33929f2cf9a3","Type":"ContainerStarted","Data":"f1869b57ee704bc529dc264795865c3f872df7083eb41a0df5814c02ae0f142a"} Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.275428 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-j5pxc" podStartSLOduration=121.275409934 podStartE2EDuration="2m1.275409934s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:23.274704762 +0000 UTC m=+140.807218683" watchObservedRunningTime="2025-11-28 11:24:23.275409934 +0000 UTC m=+140.807923855" Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.365172 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.365473 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.865458458 +0000 UTC m=+141.397972379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.465854 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.469134 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:23.969083634 +0000 UTC m=+141.501597565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.570885 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-8wmwz" podStartSLOduration=121.570866541 podStartE2EDuration="2m1.570866541s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:23.552803116 +0000 UTC m=+141.085317027" watchObservedRunningTime="2025-11-28 11:24:23.570866541 +0000 UTC m=+141.103380462" Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.571425 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:24.071414518 +0000 UTC m=+141.603928439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.571715 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lhslh" podStartSLOduration=122.571710837 podStartE2EDuration="2m2.571710837s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:23.509720266 +0000 UTC m=+141.042234187" watchObservedRunningTime="2025-11-28 11:24:23.571710837 +0000 UTC m=+141.104224758" Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.571197 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.635257 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" podStartSLOduration=121.635240918 podStartE2EDuration="2m1.635240918s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:23.626716297 +0000 UTC m=+141.159230208" watchObservedRunningTime="2025-11-28 11:24:23.635240918 +0000 UTC m=+141.167754839" Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.677677 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.677984 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:24.177968537 +0000 UTC m=+141.710482458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.782796 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.783467 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:24.283454452 +0000 UTC m=+141.815968373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.846957 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2"] Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.886493 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" podStartSLOduration=122.886471818 podStartE2EDuration="2m2.886471818s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:23.850614158 +0000 UTC m=+141.383128069" watchObservedRunningTime="2025-11-28 11:24:23.886471818 +0000 UTC m=+141.418985739" Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.887053 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.887356 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:24.387341786 +0000 UTC m=+141.919855707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.913919 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-s926b"] Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.925040 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd"] Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.950316 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-5hjfs" podStartSLOduration=121.950298948 podStartE2EDuration="2m1.950298948s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:23.935563989 +0000 UTC m=+141.468077910" watchObservedRunningTime="2025-11-28 11:24:23.950298948 +0000 UTC m=+141.482812869" Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.951517 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn"] Nov 28 11:24:23 crc kubenswrapper[4862]: I1128 11:24:23.988204 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:23 crc kubenswrapper[4862]: E1128 11:24:23.988545 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:24.488529224 +0000 UTC m=+142.021043145 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.097356 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.097725 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:24.597701246 +0000 UTC m=+142.130215167 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.141980 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tx94r" podStartSLOduration=122.141952603 podStartE2EDuration="2m2.141952603s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:23.960275075 +0000 UTC m=+141.492788996" watchObservedRunningTime="2025-11-28 11:24:24.141952603 +0000 UTC m=+141.674466524" Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.153250 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kpm6d"] Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.200459 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.201002 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:24.700984401 +0000 UTC m=+142.233498322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.273632 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk"] Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.301273 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.302835 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:24.802817469 +0000 UTC m=+142.335331390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.323972 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" event={"ID":"47f57404-6c91-49f7-bfc5-8a4c2ddaca35","Type":"ContainerStarted","Data":"0e8f9e89752db60309a0ef1ab8cc4748e513ce2361f3929ea5a66b54f89547a3"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.327628 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" event={"ID":"67f7715f-3bb2-4b68-9f0c-fda6aee68678","Type":"ContainerStarted","Data":"d394719eba8646b725dd0ac705e7acf82f7bd45d9cbcc5121b782149670d4250"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.341469 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pzshw" event={"ID":"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7","Type":"ContainerStarted","Data":"34f3e7a06e9df9bdfc47fd83388550a63944b3f82f342cda02971eda7c96f379"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.361368 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-l8mhs"] Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.364162 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9"] Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.365895 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw"] Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.366034 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" event={"ID":"33f176fd-b230-4a6c-993f-68ff9a8e3cd7","Type":"ContainerStarted","Data":"b1e08d37c91d75e6a2dbb7d5cc824559044ba6dacadc573510300735ca093f3b"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.384808 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-pzshw" podStartSLOduration=122.384787466 podStartE2EDuration="2m2.384787466s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:24.384704534 +0000 UTC m=+141.917218455" watchObservedRunningTime="2025-11-28 11:24:24.384787466 +0000 UTC m=+141.917301387" Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.394594 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" event={"ID":"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7","Type":"ContainerStarted","Data":"a17d9359b571fc5c183d73cdab785dd031b9efc8df6926c08f23bb959e820a67"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.405330 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.405649 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:24.905635779 +0000 UTC m=+142.438149700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.405746 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" event={"ID":"a23fb95f-e0c8-4681-b797-035639fed352","Type":"ContainerStarted","Data":"4f9ad9083ac2598c6f5a339c234e1fb61d648d8308db92877e5372dbb5605590"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.417989 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" event={"ID":"f70be661-2a14-4ba9-a271-ce98998cf14c","Type":"ContainerStarted","Data":"0abc080631abe478054cf3c3ae203bbd4d5c6133d8f920f762c6ea36ee7a7149"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.423878 4862 generic.go:334] "Generic (PLEG): container finished" podID="90054579-c9cd-44f7-8d0d-bca5c4bd6e20" containerID="ca167bec16081f53f6d074328fe7b6e489d7ba156ad2bdf118d4f5a8b04eb5b2" exitCode=0 Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.423934 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" event={"ID":"90054579-c9cd-44f7-8d0d-bca5c4bd6e20","Type":"ContainerDied","Data":"ca167bec16081f53f6d074328fe7b6e489d7ba156ad2bdf118d4f5a8b04eb5b2"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.435730 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" event={"ID":"3b577ecb-1896-4699-bd7e-33929f2cf9a3","Type":"ContainerStarted","Data":"7ce732110dabdb015cb436adb7d4492085119d3ed7379ec98d25dff4366d4872"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.436268 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.443548 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-nqpc6" event={"ID":"d15923b5-8e5e-4c97-9137-2e84e65d63c4","Type":"ContainerStarted","Data":"86882cd3ece727c6d42bc4645f2fd26ce365c0ef688f9f065bdf59365ba79e0b"} Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.445634 4862 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nbzb4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.445671 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" podUID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.473072 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q7bbb"] Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.491571 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-d5qm9"] Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.506734 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.507506 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.007481339 +0000 UTC m=+142.539995260 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.514364 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" podStartSLOduration=122.514349657 podStartE2EDuration="2m2.514349657s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:24.512778307 +0000 UTC m=+142.045292228" watchObservedRunningTime="2025-11-28 11:24:24.514349657 +0000 UTC m=+142.046863578" Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.608533 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.611697 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.111679232 +0000 UTC m=+142.644193153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.721565 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.723421 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.726279 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.726574 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.226560806 +0000 UTC m=+142.759074727 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.746050 4862 patch_prober.go:28] interesting pod/apiserver-76f77b778f-wt49p container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]log ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]etcd ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/generic-apiserver-start-informers ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/max-in-flight-filter ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 28 11:24:24 crc kubenswrapper[4862]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/project.openshift.io-projectcache ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/openshift.io-startinformers ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 28 11:24:24 crc kubenswrapper[4862]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 28 11:24:24 crc kubenswrapper[4862]: livez check failed Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.746119 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" podUID="f112a896-221b-4cfc-aef3-417a60c5be62" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.827504 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.827843 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.327831557 +0000 UTC m=+142.860345478 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.929776 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.929960 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.429928164 +0000 UTC m=+142.962442085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:24 crc kubenswrapper[4862]: I1128 11:24:24.930430 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:24 crc kubenswrapper[4862]: E1128 11:24:24.930763 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.430739419 +0000 UTC m=+142.963253340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.031499 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.032011 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.5319908 +0000 UTC m=+143.064504711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.133374 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.134030 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.634013404 +0000 UTC m=+143.166527325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.239715 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.240121 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.740079398 +0000 UTC m=+143.272593319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.343002 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.346729 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.846713619 +0000 UTC m=+143.379227530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.447312 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.448011 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:25.94799625 +0000 UTC m=+143.480510171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.492361 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" event={"ID":"8fd16cd6-9d43-4843-b091-aaf0686c9b94","Type":"ContainerStarted","Data":"017b25cba45db7d82bd9ee95f8c3d932a7474dc9671466dc24aa6e77d8acaa3e"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.492408 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" event={"ID":"8fd16cd6-9d43-4843-b091-aaf0686c9b94","Type":"ContainerStarted","Data":"da48b90eba25ae081b99c5b6943dcd4b7bd70db28715429a4f00573520280fb2"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.545437 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" event={"ID":"a23fb95f-e0c8-4681-b797-035639fed352","Type":"ContainerStarted","Data":"7b65944ff94fc2d55b2d50cfbd4cbb107e982b2bd8e1741f0b542e47bf4bc1ff"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.546302 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.552054 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.553725 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:26.053709062 +0000 UTC m=+143.586222983 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.560309 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" event={"ID":"3f40b08d-cf89-4ff6-a766-a38aab93f8dd","Type":"ContainerStarted","Data":"bb77d6ea3298ffb3612404bc84811f2a71c39ec69d31150e2a2bbffa624d83a6"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.569739 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.600556 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" event={"ID":"3a1a2178-787b-4af4-a15f-225ec25b4216","Type":"ContainerStarted","Data":"c16b30d0e98a6155300c98aee0f1b066962f5ae8d8f18da1c469adb54c3c682a"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.630260 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" event={"ID":"170cfb20-5cf0-4050-944b-6b17b0e2decc","Type":"ContainerStarted","Data":"52c6f140280da0d467f28a3ad91c4a540f19a78fc6d649a58609458956c82690"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.650378 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" event={"ID":"47f57404-6c91-49f7-bfc5-8a4c2ddaca35","Type":"ContainerStarted","Data":"4635ec981bff02d8a754541c7018fb06a40274fbddbd57713a8f50a88e4fceea"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.654991 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.656425 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:26.156403798 +0000 UTC m=+143.688917719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.679203 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xw6qk" podStartSLOduration=123.679179712 podStartE2EDuration="2m3.679179712s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:25.677755977 +0000 UTC m=+143.210269898" watchObservedRunningTime="2025-11-28 11:24:25.679179712 +0000 UTC m=+143.211693633" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.680256 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m6kbw" podStartSLOduration=123.680251136 podStartE2EDuration="2m3.680251136s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:25.588947473 +0000 UTC m=+143.121461394" watchObservedRunningTime="2025-11-28 11:24:25.680251136 +0000 UTC m=+143.212765057" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.690124 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" event={"ID":"9b3ee9cc-2155-4e40-a8e3-6ed95db6962b","Type":"ContainerStarted","Data":"d490deb76401b3198f7cf8f6a91cae97361338a191ac5441342529e6cf5a86df"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.742258 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2bhn7" event={"ID":"269efc51-3dfb-4888-bc8f-4f681ac5d2c6","Type":"ContainerStarted","Data":"3f4c0f143bc5bcb6ebaba008cc2c4b4ca412320dd1dbb125786f81b52519fd79"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.762720 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.763636 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:26.263611728 +0000 UTC m=+143.796125649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.764419 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" event={"ID":"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af","Type":"ContainerStarted","Data":"9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.765036 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.798877 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.804672 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9xwmh" podStartSLOduration=123.804647193 podStartE2EDuration="2m3.804647193s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:25.800683467 +0000 UTC m=+143.333197388" watchObservedRunningTime="2025-11-28 11:24:25.804647193 +0000 UTC m=+143.337161124" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.805162 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-45tfc" event={"ID":"faca5cde-65a6-429e-816c-a97fa57c5bf1","Type":"ContainerStarted","Data":"c3919aec86229f42fc79447f8016929c5cc71931786d2641baaed94cbdf419c0"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.805997 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-45tfc" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.806136 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w9lkd" podStartSLOduration=123.80613126 podStartE2EDuration="2m3.80613126s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:25.716382776 +0000 UTC m=+143.248896697" watchObservedRunningTime="2025-11-28 11:24:25.80613126 +0000 UTC m=+143.338645181" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.816925 4862 patch_prober.go:28] interesting pod/downloads-7954f5f757-45tfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.816994 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-45tfc" podUID="faca5cde-65a6-429e-816c-a97fa57c5bf1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.817194 4862 patch_prober.go:28] interesting pod/router-default-5444994796-2bhn7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 11:24:25 crc kubenswrapper[4862]: [-]has-synced failed: reason withheld Nov 28 11:24:25 crc kubenswrapper[4862]: [+]process-running ok Nov 28 11:24:25 crc kubenswrapper[4862]: healthz check failed Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.817217 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2bhn7" podUID="269efc51-3dfb-4888-bc8f-4f681ac5d2c6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.841644 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" event={"ID":"d14c0311-e7c1-4464-8872-f5e28b3ee9a2","Type":"ContainerStarted","Data":"fae613cf8908d6080e0124d2a2d1afc7fae8dd751d6f0b5716b2ec5edf35ebde"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.852948 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" podStartSLOduration=124.852934158 podStartE2EDuration="2m4.852934158s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:25.851885055 +0000 UTC m=+143.384398976" watchObservedRunningTime="2025-11-28 11:24:25.852934158 +0000 UTC m=+143.385448079" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.864423 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.865835 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:26.365818768 +0000 UTC m=+143.898332689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.887965 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" event={"ID":"5c0e4c02-c077-4c5e-bc39-0601250102b3","Type":"ContainerStarted","Data":"252a1ed6555e1921e717219cbeb97353aff729aac96a244182d97145aaa9e61c"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.897148 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" event={"ID":"67f7715f-3bb2-4b68-9f0c-fda6aee68678","Type":"ContainerStarted","Data":"7f0d856dc8560239d9b465c983fe1bf2abe0fc3c26ba0f297b4fd08dd9e16b7a"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.931681 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-nqpc6" event={"ID":"d15923b5-8e5e-4c97-9137-2e84e65d63c4","Type":"ContainerStarted","Data":"bf95c790931d6b174cdaab9ec6956936f81a2776ce773efcb2038a4a42a87fb8"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.933743 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-2bhn7" podStartSLOduration=123.933714708 podStartE2EDuration="2m3.933714708s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:25.932864431 +0000 UTC m=+143.465378342" watchObservedRunningTime="2025-11-28 11:24:25.933714708 +0000 UTC m=+143.466228629" Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.948689 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kpm6d" event={"ID":"80fef19a-293d-41fa-9ece-82f411f960a0","Type":"ContainerStarted","Data":"80ead9805f74449847d8ce82eb11e5e809a91b0e459056eb816f9e654f316452"} Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.968524 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:25 crc kubenswrapper[4862]: E1128 11:24:25.971012 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:26.470981853 +0000 UTC m=+144.003495774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:25 crc kubenswrapper[4862]: I1128 11:24:25.986947 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" event={"ID":"ce5a45b8-754f-46aa-a0eb-8325aa5a3868","Type":"ContainerStarted","Data":"ed03c57043560714c1531228c1f122201aefc637d20264860d96ffee48e22e7e"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.015601 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" event={"ID":"f70be661-2a14-4ba9-a271-ce98998cf14c","Type":"ContainerStarted","Data":"e9ae4507ebe4757dcb41289ad39a6dcb3b93cfc7f746a2b9f4e48b2674373fae"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.016570 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.030050 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-45tfc" podStartSLOduration=124.0300264 podStartE2EDuration="2m4.0300264s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:25.986462415 +0000 UTC m=+143.518976336" watchObservedRunningTime="2025-11-28 11:24:26.0300264 +0000 UTC m=+143.562540321" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.031129 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" podStartSLOduration=125.031124145 podStartE2EDuration="2m5.031124145s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.028491801 +0000 UTC m=+143.561005742" watchObservedRunningTime="2025-11-28 11:24:26.031124145 +0000 UTC m=+143.563638056" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.036005 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" event={"ID":"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed","Type":"ContainerStarted","Data":"e06fba3bac9c789b7a250263274415d2923b4377d7d65a1816c020501690a271"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.036071 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" event={"ID":"cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed","Type":"ContainerStarted","Data":"706e1e1cc8b3aeb3f02fe18dfaf87047510f33782a95ff6dede6e5a4ea8b85ed"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.036896 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.061604 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" event={"ID":"ee1f3625-a084-4e0d-99af-627305b0f8f2","Type":"ContainerStarted","Data":"13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.062654 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.063907 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" event={"ID":"a4038318-be77-4a50-a6f2-611308bda6e7","Type":"ContainerStarted","Data":"f226d5020a8cf35ab2e1ae3c71c8d1dec7d79af775819d1552a6f2ac1080e930"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.064364 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.072573 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.072948 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" podStartSLOduration=124.072936514 podStartE2EDuration="2m4.072936514s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.071511579 +0000 UTC m=+143.604025490" watchObservedRunningTime="2025-11-28 11:24:26.072936514 +0000 UTC m=+143.605450435" Nov 28 11:24:26 crc kubenswrapper[4862]: E1128 11:24:26.074057 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:26.57404339 +0000 UTC m=+144.106557311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.074571 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.091674 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" event={"ID":"b2656085-4bdf-495b-a4e7-92042752c35d","Type":"ContainerStarted","Data":"29bc6996b73340b5500ee17a83b9c0d297a2da5919892f8e333f35eb56edabc1"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.094820 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.100269 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-d5qm9" event={"ID":"bc932a43-68f4-4582-8894-6fcae516668c","Type":"ContainerStarted","Data":"a26840d270c9bd389dc7dd9571e9a08e9d8368284c4bb133fd6952c5148e7c91"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.102407 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" podStartSLOduration=124.102394501 podStartE2EDuration="2m4.102394501s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.101458952 +0000 UTC m=+143.633972883" watchObservedRunningTime="2025-11-28 11:24:26.102394501 +0000 UTC m=+143.634908422" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.119991 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" event={"ID":"daf7469b-a0bd-4682-a6df-502827ae1306","Type":"ContainerStarted","Data":"f10f0d77c8507add840938d818168e3d9f36c39e374259c86822d4e97e5a3cfc"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.120405 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" event={"ID":"daf7469b-a0bd-4682-a6df-502827ae1306","Type":"ContainerStarted","Data":"98c96003322f1b8e1dbb1e5fce013cac57334440a16906f6244d65398ea731be"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.169870 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" event={"ID":"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7","Type":"ContainerStarted","Data":"94d14e567d7579f3290ff3680f0db4b2eddb9e291f0dc006bf7af2dde1e3cb44"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.190551 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:26 crc kubenswrapper[4862]: E1128 11:24:26.191284 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:26.691269238 +0000 UTC m=+144.223783159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.203408 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" event={"ID":"f62f275b-b840-4b88-8992-a2f6ecc08ab0","Type":"ContainerStarted","Data":"dd4d4128b1847e7a6cfd370135043345da3b240acfacb4a4a0ad05b09a24df9f"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.203466 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" event={"ID":"f62f275b-b840-4b88-8992-a2f6ecc08ab0","Type":"ContainerStarted","Data":"46f9283261a5a3aadca7b5ddc7625ec0cae2dcf91ade85e7d15fc2ad5f9ea429"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.203674 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gvpct" podStartSLOduration=124.20361386 podStartE2EDuration="2m4.20361386s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.129769792 +0000 UTC m=+143.662283713" watchObservedRunningTime="2025-11-28 11:24:26.20361386 +0000 UTC m=+143.736127781" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.210831 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" event={"ID":"33f176fd-b230-4a6c-993f-68ff9a8e3cd7","Type":"ContainerStarted","Data":"500fa0ab47d98c436d9e32eb3b4d081ebd7c98f07ac97161f5d99a612ee5c250"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.254161 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" event={"ID":"c34ee8a3-cf6f-4d17-9638-8950c56b9fc8","Type":"ContainerStarted","Data":"04cf39f5314b2814cd19366712f024e587463d77c7ee6cb18808646de00944d4"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.254202 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" event={"ID":"c34ee8a3-cf6f-4d17-9638-8950c56b9fc8","Type":"ContainerStarted","Data":"141db47d552d82e43e2dc667cb67da878e535bbe425efefe7636bfe02285d0fe"} Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.256526 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-nqpc6" podStartSLOduration=8.256515063 podStartE2EDuration="8.256515063s" podCreationTimestamp="2025-11-28 11:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.189016036 +0000 UTC m=+143.721529957" watchObservedRunningTime="2025-11-28 11:24:26.256515063 +0000 UTC m=+143.789028984" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.264693 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s86df" podStartSLOduration=124.264676452 podStartE2EDuration="2m4.264676452s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.22717939 +0000 UTC m=+143.759693321" watchObservedRunningTime="2025-11-28 11:24:26.264676452 +0000 UTC m=+143.797190373" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.269303 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.297908 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:26 crc kubenswrapper[4862]: E1128 11:24:26.299598 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:26.799573082 +0000 UTC m=+144.332087003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.364442 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-d5qm9" podStartSLOduration=8.364420984 podStartE2EDuration="8.364420984s" podCreationTimestamp="2025-11-28 11:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.293248461 +0000 UTC m=+143.825762382" watchObservedRunningTime="2025-11-28 11:24:26.364420984 +0000 UTC m=+143.896934905" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.365407 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" podStartSLOduration=124.365402096 podStartE2EDuration="2m4.365402096s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.363277558 +0000 UTC m=+143.895791489" watchObservedRunningTime="2025-11-28 11:24:26.365402096 +0000 UTC m=+143.897916007" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.407378 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:26 crc kubenswrapper[4862]: E1128 11:24:26.407682 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:26.90766905 +0000 UTC m=+144.440182971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.440387 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" podStartSLOduration=124.44037023 podStartE2EDuration="2m4.44037023s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.394306735 +0000 UTC m=+143.926820656" watchObservedRunningTime="2025-11-28 11:24:26.44037023 +0000 UTC m=+143.972884151" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.440793 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-q7bbb" podStartSLOduration=124.440787163 podStartE2EDuration="2m4.440787163s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.440648579 +0000 UTC m=+143.973162500" watchObservedRunningTime="2025-11-28 11:24:26.440787163 +0000 UTC m=+143.973301084" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.508469 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:26 crc kubenswrapper[4862]: E1128 11:24:26.509364 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.009321873 +0000 UTC m=+144.541835794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.542363 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" podStartSLOduration=124.542346943 podStartE2EDuration="2m4.542346943s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.494366307 +0000 UTC m=+144.026880228" watchObservedRunningTime="2025-11-28 11:24:26.542346943 +0000 UTC m=+144.074860864" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.548004 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.570253 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5gcl2" podStartSLOduration=124.57023492 podStartE2EDuration="2m4.57023492s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.543565922 +0000 UTC m=+144.076079843" watchObservedRunningTime="2025-11-28 11:24:26.57023492 +0000 UTC m=+144.102748831" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.597854 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7jpvc" podStartSLOduration=124.597838698 podStartE2EDuration="2m4.597838698s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.572010537 +0000 UTC m=+144.104524458" watchObservedRunningTime="2025-11-28 11:24:26.597838698 +0000 UTC m=+144.130352619" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.615744 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:26 crc kubenswrapper[4862]: E1128 11:24:26.616038 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.116027037 +0000 UTC m=+144.648540958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.659664 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" podStartSLOduration=124.659644094 podStartE2EDuration="2m4.659644094s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.597543839 +0000 UTC m=+144.130057760" watchObservedRunningTime="2025-11-28 11:24:26.659644094 +0000 UTC m=+144.192158015" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.716562 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:26 crc kubenswrapper[4862]: E1128 11:24:26.716965 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.216952286 +0000 UTC m=+144.749466207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.745731 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-qlsvq" podStartSLOduration=124.745711491 podStartE2EDuration="2m4.745711491s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:26.744466821 +0000 UTC m=+144.276980732" watchObservedRunningTime="2025-11-28 11:24:26.745711491 +0000 UTC m=+144.278225412" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.747071 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wbstd"] Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.747940 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.760404 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.804261 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wbstd"] Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.819249 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.819304 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-utilities\") pod \"community-operators-wbstd\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.819341 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-catalog-content\") pod \"community-operators-wbstd\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.819364 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfzcg\" (UniqueName: \"kubernetes.io/projected/2f359e91-8745-46c0-acad-6bc9795f04ba-kube-api-access-wfzcg\") pod \"community-operators-wbstd\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: E1128 11:24:26.819669 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.319658103 +0000 UTC m=+144.852172024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.820243 4862 patch_prober.go:28] interesting pod/router-default-5444994796-2bhn7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 11:24:26 crc kubenswrapper[4862]: [-]has-synced failed: reason withheld Nov 28 11:24:26 crc kubenswrapper[4862]: [+]process-running ok Nov 28 11:24:26 crc kubenswrapper[4862]: healthz check failed Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.820291 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2bhn7" podUID="269efc51-3dfb-4888-bc8f-4f681ac5d2c6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.920560 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j5c55"] Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.920728 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.920973 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-utilities\") pod \"community-operators-wbstd\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.921022 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-catalog-content\") pod \"community-operators-wbstd\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.921045 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfzcg\" (UniqueName: \"kubernetes.io/projected/2f359e91-8745-46c0-acad-6bc9795f04ba-kube-api-access-wfzcg\") pod \"community-operators-wbstd\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: E1128 11:24:26.921469 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.42145398 +0000 UTC m=+144.953967901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.921975 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-catalog-content\") pod \"community-operators-wbstd\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.922164 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-utilities\") pod \"community-operators-wbstd\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.923130 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.932561 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.965458 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfzcg\" (UniqueName: \"kubernetes.io/projected/2f359e91-8745-46c0-acad-6bc9795f04ba-kube-api-access-wfzcg\") pod \"community-operators-wbstd\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:26 crc kubenswrapper[4862]: I1128 11:24:26.990425 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j5c55"] Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.030770 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-catalog-content\") pod \"certified-operators-j5c55\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.030833 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-utilities\") pod \"certified-operators-j5c55\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.030858 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.030898 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvgdr\" (UniqueName: \"kubernetes.io/projected/19d8710e-5f52-452e-8b13-854546736587-kube-api-access-tvgdr\") pod \"certified-operators-j5c55\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.031253 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.531241952 +0000 UTC m=+145.063755873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.041262 4862 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-kl4jk container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.041311 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" podUID="cec956a2-9b41-4d4e-a0f7-d2e1dbd4f5ed" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.067490 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.123750 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4hzfd"] Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.125384 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.132200 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.132460 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.63244509 +0000 UTC m=+145.164959011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.132494 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvgdr\" (UniqueName: \"kubernetes.io/projected/19d8710e-5f52-452e-8b13-854546736587-kube-api-access-tvgdr\") pod \"certified-operators-j5c55\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.132582 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-catalog-content\") pod \"certified-operators-j5c55\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.132611 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-utilities\") pod \"certified-operators-j5c55\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.132637 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.132961 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.632951716 +0000 UTC m=+145.165465647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.133608 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-catalog-content\") pod \"certified-operators-j5c55\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.135337 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-utilities\") pod \"certified-operators-j5c55\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.159654 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hzfd"] Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.165170 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvgdr\" (UniqueName: \"kubernetes.io/projected/19d8710e-5f52-452e-8b13-854546736587-kube-api-access-tvgdr\") pod \"certified-operators-j5c55\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.243588 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.260896 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.261547 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-utilities\") pod \"community-operators-4hzfd\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.261670 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz9jc\" (UniqueName: \"kubernetes.io/projected/6edecb40-afd9-40ab-8f5b-227730410e66-kube-api-access-tz9jc\") pod \"community-operators-4hzfd\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.261789 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-catalog-content\") pod \"community-operators-4hzfd\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.262461 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.762445045 +0000 UTC m=+145.294958966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.305687 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-d5qm9" event={"ID":"bc932a43-68f4-4582-8894-6fcae516668c","Type":"ContainerStarted","Data":"485b3358efd9d446edcab2b82987032f712b54150273985f9d868f2a4d03ba93"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.307964 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" event={"ID":"9b3ee9cc-2155-4e40-a8e3-6ed95db6962b","Type":"ContainerStarted","Data":"ff0c1c7cad2c825f0c7dc86b6256caa59a1845d875ffa19160de7884b114d590"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.309874 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" event={"ID":"f70be661-2a14-4ba9-a271-ce98998cf14c","Type":"ContainerStarted","Data":"4f1936d537ec2c0fdafd9dcc02d961318482fe309e9d00c62f484f166587a337"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.329792 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cph87"] Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.331179 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.336885 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" event={"ID":"d14c0311-e7c1-4464-8872-f5e28b3ee9a2","Type":"ContainerStarted","Data":"d6712ac106c19d6ad10600fbc74ee3eb3b53e2b68fec040adb685f31c0dcc487"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.359168 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bp9bf" event={"ID":"5c0e4c02-c077-4c5e-bc39-0601250102b3","Type":"ContainerStarted","Data":"14ea97d046660fab82eaac2434778a2b19bda6162efcdc2ebc77354748c6aff3"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.364511 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cph87"] Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.365330 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.365408 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-utilities\") pod \"community-operators-4hzfd\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.365464 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz9jc\" (UniqueName: \"kubernetes.io/projected/6edecb40-afd9-40ab-8f5b-227730410e66-kube-api-access-tz9jc\") pod \"community-operators-4hzfd\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.365489 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-catalog-content\") pod \"community-operators-4hzfd\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.367006 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.86699314 +0000 UTC m=+145.399507061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.367449 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-utilities\") pod \"community-operators-4hzfd\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.370289 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9f6sc" podStartSLOduration=125.370262004 podStartE2EDuration="2m5.370262004s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:27.345958121 +0000 UTC m=+144.878472042" watchObservedRunningTime="2025-11-28 11:24:27.370262004 +0000 UTC m=+144.902775925" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.372574 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-catalog-content\") pod \"community-operators-4hzfd\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.389114 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgpc4" podStartSLOduration=126.389085742 podStartE2EDuration="2m6.389085742s" podCreationTimestamp="2025-11-28 11:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:27.388435492 +0000 UTC m=+144.920949413" watchObservedRunningTime="2025-11-28 11:24:27.389085742 +0000 UTC m=+144.921599663" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.400386 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" event={"ID":"170cfb20-5cf0-4050-944b-6b17b0e2decc","Type":"ContainerStarted","Data":"3bcc6b352b113dcbe005a37d6aa0d82970192782a180dc7ef992b39e0d78c7de"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.450686 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz9jc\" (UniqueName: \"kubernetes.io/projected/6edecb40-afd9-40ab-8f5b-227730410e66-kube-api-access-tz9jc\") pod \"community-operators-4hzfd\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.451560 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" event={"ID":"90054579-c9cd-44f7-8d0d-bca5c4bd6e20","Type":"ContainerStarted","Data":"cbf43f427e95f4b5177d75b5c754bce85560926a2670e5485cedaaa6b7f6528a"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.454548 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.454627 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-7mhnb" podStartSLOduration=125.454598126 podStartE2EDuration="2m5.454598126s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:27.452442927 +0000 UTC m=+144.984956848" watchObservedRunningTime="2025-11-28 11:24:27.454598126 +0000 UTC m=+144.987112037" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.468632 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.469146 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-catalog-content\") pod \"certified-operators-cph87\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.469194 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djjls\" (UniqueName: \"kubernetes.io/projected/22c7db45-0f31-4bc7-b93d-4470c51434b1-kube-api-access-djjls\") pod \"certified-operators-cph87\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.469310 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-utilities\") pod \"certified-operators-cph87\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.471622 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:27.971594586 +0000 UTC m=+145.504108507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.485790 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kpm6d" event={"ID":"80fef19a-293d-41fa-9ece-82f411f960a0","Type":"ContainerStarted","Data":"ec81e1ac6bce2c256a7de17684f5f7b0d5ba72ba5e8111e419949d5681a1cc3e"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.485872 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kpm6d" event={"ID":"80fef19a-293d-41fa-9ece-82f411f960a0","Type":"ContainerStarted","Data":"0ecbe1cd9f1d83409f25b681b2ccfeae42e4c9ad269743aa5826cd4983da8922"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.486683 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.488081 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" event={"ID":"b2656085-4bdf-495b-a4e7-92042752c35d","Type":"ContainerStarted","Data":"f3b676e3d85aee799d2d4a922547422c91075f8e8ee2001fbabfa34def2fb5e0"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.494111 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" podStartSLOduration=125.494084292 podStartE2EDuration="2m5.494084292s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:27.49056294 +0000 UTC m=+145.023076851" watchObservedRunningTime="2025-11-28 11:24:27.494084292 +0000 UTC m=+145.026598213" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.523823 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8hqk9" event={"ID":"daf7469b-a0bd-4682-a6df-502827ae1306","Type":"ContainerStarted","Data":"8d61e4720e200be238ed88f630532e2e1b6b1b36015a663f27804fe4e9947b3f"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.539957 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-kpm6d" podStartSLOduration=9.5399185 podStartE2EDuration="9.5399185s" podCreationTimestamp="2025-11-28 11:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:27.525178781 +0000 UTC m=+145.057692702" watchObservedRunningTime="2025-11-28 11:24:27.5399185 +0000 UTC m=+145.072432421" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.553419 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s926b" event={"ID":"f4da0ab8-bdd5-44eb-ac21-626d7c983fa7","Type":"ContainerStarted","Data":"e0c2777c07490a962b6d8746c142a27b9d6a57218dee8bba17462aaf1b97a477"} Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.555538 4862 patch_prober.go:28] interesting pod/downloads-7954f5f757-45tfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.555589 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-45tfc" podUID="faca5cde-65a6-429e-816c-a97fa57c5bf1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.566484 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kl4jk" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.572650 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.572702 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-utilities\") pod \"certified-operators-cph87\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.573188 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-utilities\") pod \"certified-operators-cph87\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.575301 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-catalog-content\") pod \"certified-operators-cph87\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.575379 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djjls\" (UniqueName: \"kubernetes.io/projected/22c7db45-0f31-4bc7-b93d-4470c51434b1-kube-api-access-djjls\") pod \"certified-operators-cph87\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.575857 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-catalog-content\") pod \"certified-operators-cph87\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.576941 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:28.076919636 +0000 UTC m=+145.609433787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.606219 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djjls\" (UniqueName: \"kubernetes.io/projected/22c7db45-0f31-4bc7-b93d-4470c51434b1-kube-api-access-djjls\") pod \"certified-operators-cph87\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.682964 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.685274 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:28.185251542 +0000 UTC m=+145.717765463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.688836 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.703368 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wbstd"] Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.787308 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.787999 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:28.287973369 +0000 UTC m=+145.820487290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.824300 4862 patch_prober.go:28] interesting pod/router-default-5444994796-2bhn7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 11:24:27 crc kubenswrapper[4862]: [-]has-synced failed: reason withheld Nov 28 11:24:27 crc kubenswrapper[4862]: [+]process-running ok Nov 28 11:24:27 crc kubenswrapper[4862]: healthz check failed Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.824351 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2bhn7" podUID="269efc51-3dfb-4888-bc8f-4f681ac5d2c6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.888395 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j5c55"] Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.889293 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:27 crc kubenswrapper[4862]: E1128 11:24:27.889737 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:28.389714714 +0000 UTC m=+145.922228635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.957456 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.958404 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.965537 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.969394 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.997888 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 11:24:27 crc kubenswrapper[4862]: I1128 11:24:27.999013 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:27.999586 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:28.499568268 +0000 UTC m=+146.032082189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.103398 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.103588 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.103659 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.103761 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:28.603744631 +0000 UTC m=+146.136258552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.207207 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.207253 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.207320 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.207401 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.207867 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:28.707857922 +0000 UTC m=+146.240371843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.240981 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.259392 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hzfd"] Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.268048 4862 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.289526 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cph87"] Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.307896 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.308260 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:28.808240355 +0000 UTC m=+146.340754276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.346438 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.412370 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.412726 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:28.912713848 +0000 UTC m=+146.445227769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.513892 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.514215 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.014188165 +0000 UTC m=+146.546702086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.514517 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.514813 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.014803544 +0000 UTC m=+146.547317465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.560690 4862 generic.go:334] "Generic (PLEG): container finished" podID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerID="4c8ff714cb99c1551fabc9f5393064ba17c22e985ca7522803598711941cbcee" exitCode=0 Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.560753 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbstd" event={"ID":"2f359e91-8745-46c0-acad-6bc9795f04ba","Type":"ContainerDied","Data":"4c8ff714cb99c1551fabc9f5393064ba17c22e985ca7522803598711941cbcee"} Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.560779 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbstd" event={"ID":"2f359e91-8745-46c0-acad-6bc9795f04ba","Type":"ContainerStarted","Data":"7182ba62bae339490dc42abeea09fd665f32ccbe642f17b17a18991caba0fc75"} Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.563061 4862 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.566815 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cph87" event={"ID":"22c7db45-0f31-4bc7-b93d-4470c51434b1","Type":"ContainerStarted","Data":"0a91f6cde582434aa43401b24dc91ae2a291732b32400964ab4fc090cb6075fb"} Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.569357 4862 generic.go:334] "Generic (PLEG): container finished" podID="19d8710e-5f52-452e-8b13-854546736587" containerID="013c87cac0693d27468d902f9763c86a0d8bde97f800b0f450b5f7a445ba5c43" exitCode=0 Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.569410 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5c55" event={"ID":"19d8710e-5f52-452e-8b13-854546736587","Type":"ContainerDied","Data":"013c87cac0693d27468d902f9763c86a0d8bde97f800b0f450b5f7a445ba5c43"} Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.569435 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5c55" event={"ID":"19d8710e-5f52-452e-8b13-854546736587","Type":"ContainerStarted","Data":"12204150939db276c9bca43802ce0e06722f2176e05bbea7c33bd81e7652836b"} Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.582736 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" event={"ID":"b2656085-4bdf-495b-a4e7-92042752c35d","Type":"ContainerStarted","Data":"2271eac4cdb7374586a8868919c71a2cf6e1af73fb9cf00932ac7b6a8966444d"} Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.582784 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" event={"ID":"b2656085-4bdf-495b-a4e7-92042752c35d","Type":"ContainerStarted","Data":"d66cd1513791e24f89ba6eeb460ad31c7b375816e88187612be3d4dc3dcf3c91"} Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.592541 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hzfd" event={"ID":"6edecb40-afd9-40ab-8f5b-227730410e66","Type":"ContainerStarted","Data":"fa3ebad9e7121fa201a127ee26fc500d31d0db42351e4e9e1d5c1d4ee3a46a04"} Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.594055 4862 patch_prober.go:28] interesting pod/downloads-7954f5f757-45tfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.594115 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-45tfc" podUID="faca5cde-65a6-429e-816c-a97fa57c5bf1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.616377 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.616546 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.11652273 +0000 UTC m=+146.649036651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.616688 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.616998 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.116991634 +0000 UTC m=+146.649505555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.718577 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.719031 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.218992458 +0000 UTC m=+146.751506369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.719930 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.726144 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.226117815 +0000 UTC m=+146.758631736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.781275 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.812384 4862 patch_prober.go:28] interesting pod/router-default-5444994796-2bhn7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 11:24:28 crc kubenswrapper[4862]: [-]has-synced failed: reason withheld Nov 28 11:24:28 crc kubenswrapper[4862]: [+]process-running ok Nov 28 11:24:28 crc kubenswrapper[4862]: healthz check failed Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.812451 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2bhn7" podUID="269efc51-3dfb-4888-bc8f-4f681ac5d2c6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.821023 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.821206 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.321177068 +0000 UTC m=+146.853690989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.821423 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.821833 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.321825029 +0000 UTC m=+146.854338950 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.923055 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.923276 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.423218563 +0000 UTC m=+146.955732484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:28 crc kubenswrapper[4862]: I1128 11:24:28.923521 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:28 crc kubenswrapper[4862]: E1128 11:24:28.923883 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 11:24:29.423876114 +0000 UTC m=+146.956390035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hw6dt" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.014448 4862 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-28T11:24:28.268076708Z","Handler":null,"Name":""} Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.017588 4862 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.017659 4862 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.024270 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.028860 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.115134 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l8jkt"] Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.116440 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.119521 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.135245 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.142353 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8jkt"] Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.147195 4862 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.147233 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.193721 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hw6dt\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.237686 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-utilities\") pod \"redhat-marketplace-l8jkt\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.238197 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-catalog-content\") pod \"redhat-marketplace-l8jkt\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.238285 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7cmt\" (UniqueName: \"kubernetes.io/projected/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-kube-api-access-m7cmt\") pod \"redhat-marketplace-l8jkt\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.339551 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-utilities\") pod \"redhat-marketplace-l8jkt\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.339852 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-catalog-content\") pod \"redhat-marketplace-l8jkt\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.339923 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7cmt\" (UniqueName: \"kubernetes.io/projected/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-kube-api-access-m7cmt\") pod \"redhat-marketplace-l8jkt\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.340462 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-utilities\") pod \"redhat-marketplace-l8jkt\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.340701 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-catalog-content\") pod \"redhat-marketplace-l8jkt\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.364706 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7cmt\" (UniqueName: \"kubernetes.io/projected/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-kube-api-access-m7cmt\") pod \"redhat-marketplace-l8jkt\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.441394 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.447768 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.526939 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2rcg7"] Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.532311 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.550918 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2rcg7"] Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.602831 4862 generic.go:334] "Generic (PLEG): container finished" podID="3a1a2178-787b-4af4-a15f-225ec25b4216" containerID="c16b30d0e98a6155300c98aee0f1b066962f5ae8d8f18da1c469adb54c3c682a" exitCode=0 Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.602894 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" event={"ID":"3a1a2178-787b-4af4-a15f-225ec25b4216","Type":"ContainerDied","Data":"c16b30d0e98a6155300c98aee0f1b066962f5ae8d8f18da1c469adb54c3c682a"} Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.610892 4862 generic.go:334] "Generic (PLEG): container finished" podID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerID="bda8bae1495e604ef029a0ecb45cbe5ce2d59fe254975608d8370b3d5ba1c77a" exitCode=0 Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.610958 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cph87" event={"ID":"22c7db45-0f31-4bc7-b93d-4470c51434b1","Type":"ContainerDied","Data":"bda8bae1495e604ef029a0ecb45cbe5ce2d59fe254975608d8370b3d5ba1c77a"} Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.618273 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" event={"ID":"b2656085-4bdf-495b-a4e7-92042752c35d","Type":"ContainerStarted","Data":"cd334c96cd473354a33a5b6ce601323d86e8c0e601f4a034180c510debf2623d"} Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.628372 4862 generic.go:334] "Generic (PLEG): container finished" podID="6edecb40-afd9-40ab-8f5b-227730410e66" containerID="808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca" exitCode=0 Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.628696 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hzfd" event={"ID":"6edecb40-afd9-40ab-8f5b-227730410e66","Type":"ContainerDied","Data":"808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca"} Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.651193 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppjfb\" (UniqueName: \"kubernetes.io/projected/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-kube-api-access-ppjfb\") pod \"redhat-marketplace-2rcg7\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.651284 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-utilities\") pod \"redhat-marketplace-2rcg7\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.651365 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-catalog-content\") pod \"redhat-marketplace-2rcg7\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.653648 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8b7604a9-e2aa-4d7b-b467-0d38d992ab93","Type":"ContainerStarted","Data":"ba4568aa3f036fc63ded867fb9f1e18f36cfca58895cbe5551699fe215e1891b"} Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.653679 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8b7604a9-e2aa-4d7b-b467-0d38d992ab93","Type":"ContainerStarted","Data":"e137a3c738993d7a76bf1dc99336b1f4936a51092df2791b741fee11a6a6e9a8"} Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.722972 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.732405 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-wt49p" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.740996 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-l8mhs" podStartSLOduration=11.74098124 podStartE2EDuration="11.74098124s" podCreationTimestamp="2025-11-28 11:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:29.740229766 +0000 UTC m=+147.272743687" watchObservedRunningTime="2025-11-28 11:24:29.74098124 +0000 UTC m=+147.273495161" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.754678 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppjfb\" (UniqueName: \"kubernetes.io/projected/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-kube-api-access-ppjfb\") pod \"redhat-marketplace-2rcg7\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.754784 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-utilities\") pod \"redhat-marketplace-2rcg7\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.754841 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-catalog-content\") pod \"redhat-marketplace-2rcg7\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.757682 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-utilities\") pod \"redhat-marketplace-2rcg7\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.770763 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-catalog-content\") pod \"redhat-marketplace-2rcg7\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.790870 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppjfb\" (UniqueName: \"kubernetes.io/projected/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-kube-api-access-ppjfb\") pod \"redhat-marketplace-2rcg7\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.806570 4862 patch_prober.go:28] interesting pod/router-default-5444994796-2bhn7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 11:24:29 crc kubenswrapper[4862]: [-]has-synced failed: reason withheld Nov 28 11:24:29 crc kubenswrapper[4862]: [+]process-running ok Nov 28 11:24:29 crc kubenswrapper[4862]: healthz check failed Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.806637 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2bhn7" podUID="269efc51-3dfb-4888-bc8f-4f681ac5d2c6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.816761 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.81674657 podStartE2EDuration="2.81674657s" podCreationTimestamp="2025-11-28 11:24:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:29.816033497 +0000 UTC m=+147.348547438" watchObservedRunningTime="2025-11-28 11:24:29.81674657 +0000 UTC m=+147.349260491" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.876193 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.912703 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hw6dt"] Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.951271 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w992j"] Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.952693 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.960467 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.960973 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.961030 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.961066 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.961180 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:29 crc kubenswrapper[4862]: W1128 11:24:29.961387 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e143937_4cf3_4fa0_aa09_381d8bf99f16.slice/crio-f27a46368de3a5a6898ace1485ade3e48dbe69ecb1ded058eab1519f876937b4 WatchSource:0}: Error finding container f27a46368de3a5a6898ace1485ade3e48dbe69ecb1ded058eab1519f876937b4: Status 404 returned error can't find the container with id f27a46368de3a5a6898ace1485ade3e48dbe69ecb1ded058eab1519f876937b4 Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.966682 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.969278 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.974295 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.976957 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w992j"] Nov 28 11:24:29 crc kubenswrapper[4862]: I1128 11:24:29.977271 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.018739 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8jkt"] Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.064656 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.065401 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwbpj\" (UniqueName: \"kubernetes.io/projected/02589a29-1121-48de-98a6-691d4f8c598f-kube-api-access-kwbpj\") pod \"redhat-operators-w992j\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.065506 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-catalog-content\") pod \"redhat-operators-w992j\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.065540 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-utilities\") pod \"redhat-operators-w992j\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.112212 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.122151 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c2nbn"] Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.123391 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.158321 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.168961 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwbpj\" (UniqueName: \"kubernetes.io/projected/02589a29-1121-48de-98a6-691d4f8c598f-kube-api-access-kwbpj\") pod \"redhat-operators-w992j\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.169031 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-catalog-content\") pod \"redhat-operators-w992j\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.169055 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-utilities\") pod \"redhat-operators-w992j\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.170168 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-utilities\") pod \"redhat-operators-w992j\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.170756 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-catalog-content\") pod \"redhat-operators-w992j\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.191981 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c2nbn"] Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.231505 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwbpj\" (UniqueName: \"kubernetes.io/projected/02589a29-1121-48de-98a6-691d4f8c598f-kube-api-access-kwbpj\") pod \"redhat-operators-w992j\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.270275 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-utilities\") pod \"redhat-operators-c2nbn\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.270322 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nszl\" (UniqueName: \"kubernetes.io/projected/e9947d52-a469-4ca3-b6b5-9482bf691c4c-kube-api-access-6nszl\") pod \"redhat-operators-c2nbn\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.270354 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-catalog-content\") pod \"redhat-operators-c2nbn\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.352894 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.376121 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-catalog-content\") pod \"redhat-operators-c2nbn\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.376283 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-utilities\") pod \"redhat-operators-c2nbn\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.376324 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nszl\" (UniqueName: \"kubernetes.io/projected/e9947d52-a469-4ca3-b6b5-9482bf691c4c-kube-api-access-6nszl\") pod \"redhat-operators-c2nbn\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.377133 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-catalog-content\") pod \"redhat-operators-c2nbn\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.377251 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-utilities\") pod \"redhat-operators-c2nbn\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.403511 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nszl\" (UniqueName: \"kubernetes.io/projected/e9947d52-a469-4ca3-b6b5-9482bf691c4c-kube-api-access-6nszl\") pod \"redhat-operators-c2nbn\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.442143 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.479472 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2rcg7"] Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.586285 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.587223 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.606898 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.606994 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.612284 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.684920 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e8740934-8747-4cea-85f9-8e127089df8e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e8740934-8747-4cea-85f9-8e127089df8e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.685007 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e8740934-8747-4cea-85f9-8e127089df8e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e8740934-8747-4cea-85f9-8e127089df8e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.699642 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2rcg7" event={"ID":"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4","Type":"ContainerStarted","Data":"5e38ae585a6868cc47c3462c63721b500a0c0717daac8d8cb8fb75d600c8d13f"} Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.718305 4862 generic.go:334] "Generic (PLEG): container finished" podID="8b7604a9-e2aa-4d7b-b467-0d38d992ab93" containerID="ba4568aa3f036fc63ded867fb9f1e18f36cfca58895cbe5551699fe215e1891b" exitCode=0 Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.718384 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8b7604a9-e2aa-4d7b-b467-0d38d992ab93","Type":"ContainerDied","Data":"ba4568aa3f036fc63ded867fb9f1e18f36cfca58895cbe5551699fe215e1891b"} Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.768638 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" event={"ID":"6e143937-4cf3-4fa0-aa09-381d8bf99f16","Type":"ContainerStarted","Data":"4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7"} Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.769136 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" event={"ID":"6e143937-4cf3-4fa0-aa09-381d8bf99f16","Type":"ContainerStarted","Data":"f27a46368de3a5a6898ace1485ade3e48dbe69ecb1ded058eab1519f876937b4"} Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.770045 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.786108 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e8740934-8747-4cea-85f9-8e127089df8e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e8740934-8747-4cea-85f9-8e127089df8e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.786170 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e8740934-8747-4cea-85f9-8e127089df8e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e8740934-8747-4cea-85f9-8e127089df8e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.786301 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e8740934-8747-4cea-85f9-8e127089df8e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e8740934-8747-4cea-85f9-8e127089df8e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.804207 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8jkt" event={"ID":"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf","Type":"ContainerStarted","Data":"d46a19c45f4f0b1f281eca375e9ad5642dd44e0dab93d230b246541dd744c2b6"} Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.822355 4862 patch_prober.go:28] interesting pod/router-default-5444994796-2bhn7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 11:24:30 crc kubenswrapper[4862]: [-]has-synced failed: reason withheld Nov 28 11:24:30 crc kubenswrapper[4862]: [+]process-running ok Nov 28 11:24:30 crc kubenswrapper[4862]: healthz check failed Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.822437 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2bhn7" podUID="269efc51-3dfb-4888-bc8f-4f681ac5d2c6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.829954 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e8740934-8747-4cea-85f9-8e127089df8e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e8740934-8747-4cea-85f9-8e127089df8e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.863655 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" podStartSLOduration=128.863632845 podStartE2EDuration="2m8.863632845s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:24:30.861874839 +0000 UTC m=+148.394388760" watchObservedRunningTime="2025-11-28 11:24:30.863632845 +0000 UTC m=+148.396146766" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.921407 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.927546 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.951673 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.952172 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.971009 4862 patch_prober.go:28] interesting pod/console-f9d7485db-pzshw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 28 11:24:30 crc kubenswrapper[4862]: I1128 11:24:30.971108 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pzshw" podUID="b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.125660 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.125686 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.136369 4862 patch_prober.go:28] interesting pod/downloads-7954f5f757-45tfc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.136406 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-45tfc" podUID="faca5cde-65a6-429e-816c-a97fa57c5bf1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.139698 4862 patch_prober.go:28] interesting pod/downloads-7954f5f757-45tfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.139725 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-45tfc" podUID="faca5cde-65a6-429e-816c-a97fa57c5bf1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.157396 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.215680 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c2nbn"] Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.345679 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w992j"] Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.375615 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 11:24:31 crc kubenswrapper[4862]: W1128 11:24:31.382282 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9947d52_a469_4ca3_b6b5_9482bf691c4c.slice/crio-9fb91a6d48b32698e42d33e5175c898a18eb9bb5859f63366b87c744963758f0 WatchSource:0}: Error finding container 9fb91a6d48b32698e42d33e5175c898a18eb9bb5859f63366b87c744963758f0: Status 404 returned error can't find the container with id 9fb91a6d48b32698e42d33e5175c898a18eb9bb5859f63366b87c744963758f0 Nov 28 11:24:31 crc kubenswrapper[4862]: W1128 11:24:31.392368 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02589a29_1121_48de_98a6_691d4f8c598f.slice/crio-71cae4f273e0d20052b93ac8e4efd9927c7637721bab89e8ba1982c1f655186e WatchSource:0}: Error finding container 71cae4f273e0d20052b93ac8e4efd9927c7637721bab89e8ba1982c1f655186e: Status 404 returned error can't find the container with id 71cae4f273e0d20052b93ac8e4efd9927c7637721bab89e8ba1982c1f655186e Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.420505 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.495654 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll696\" (UniqueName: \"kubernetes.io/projected/3a1a2178-787b-4af4-a15f-225ec25b4216-kube-api-access-ll696\") pod \"3a1a2178-787b-4af4-a15f-225ec25b4216\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.495833 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a1a2178-787b-4af4-a15f-225ec25b4216-config-volume\") pod \"3a1a2178-787b-4af4-a15f-225ec25b4216\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.496013 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a1a2178-787b-4af4-a15f-225ec25b4216-secret-volume\") pod \"3a1a2178-787b-4af4-a15f-225ec25b4216\" (UID: \"3a1a2178-787b-4af4-a15f-225ec25b4216\") " Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.498054 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a1a2178-787b-4af4-a15f-225ec25b4216-config-volume" (OuterVolumeSpecName: "config-volume") pod "3a1a2178-787b-4af4-a15f-225ec25b4216" (UID: "3a1a2178-787b-4af4-a15f-225ec25b4216"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.505378 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a1a2178-787b-4af4-a15f-225ec25b4216-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3a1a2178-787b-4af4-a15f-225ec25b4216" (UID: "3a1a2178-787b-4af4-a15f-225ec25b4216"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.508260 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a1a2178-787b-4af4-a15f-225ec25b4216-kube-api-access-ll696" (OuterVolumeSpecName: "kube-api-access-ll696") pod "3a1a2178-787b-4af4-a15f-225ec25b4216" (UID: "3a1a2178-787b-4af4-a15f-225ec25b4216"). InnerVolumeSpecName "kube-api-access-ll696". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.600831 4862 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a1a2178-787b-4af4-a15f-225ec25b4216-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.601203 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll696\" (UniqueName: \"kubernetes.io/projected/3a1a2178-787b-4af4-a15f-225ec25b4216-kube-api-access-ll696\") on node \"crc\" DevicePath \"\"" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.601214 4862 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a1a2178-787b-4af4-a15f-225ec25b4216-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.801249 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.803273 4862 patch_prober.go:28] interesting pod/router-default-5444994796-2bhn7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 11:24:31 crc kubenswrapper[4862]: [-]has-synced failed: reason withheld Nov 28 11:24:31 crc kubenswrapper[4862]: [+]process-running ok Nov 28 11:24:31 crc kubenswrapper[4862]: healthz check failed Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.803321 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2bhn7" podUID="269efc51-3dfb-4888-bc8f-4f681ac5d2c6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.825428 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f08403bfab14a8bda817629bce5965be3d92cc2f9963c2aed3a939282fe4f33e"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.825511 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"db5d1f5807ca3047fa29bc0326afcca62043a24149efb8f527026cd4322c3fc0"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.857957 4862 generic.go:334] "Generic (PLEG): container finished" podID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerID="6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06" exitCode=0 Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.858053 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8jkt" event={"ID":"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf","Type":"ContainerDied","Data":"6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.869631 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2nbn" event={"ID":"e9947d52-a469-4ca3-b6b5-9482bf691c4c","Type":"ContainerStarted","Data":"5df8f6e3fbb3eab241eba05fb3fabe95469a8d4aca35bd26faf60007749be510"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.869681 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2nbn" event={"ID":"e9947d52-a469-4ca3-b6b5-9482bf691c4c","Type":"ContainerStarted","Data":"9fb91a6d48b32698e42d33e5175c898a18eb9bb5859f63366b87c744963758f0"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.900499 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w992j" event={"ID":"02589a29-1121-48de-98a6-691d4f8c598f","Type":"ContainerStarted","Data":"71cae4f273e0d20052b93ac8e4efd9927c7637721bab89e8ba1982c1f655186e"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.909051 4862 generic.go:334] "Generic (PLEG): container finished" podID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerID="39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf" exitCode=0 Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.909273 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2rcg7" event={"ID":"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4","Type":"ContainerDied","Data":"39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.922474 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0a3d21041c5fc9db706f7247ae42de6b36d54545b61b142dcd8a9ed649e39938"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.922542 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"317201dd7deb4fc84bb40dfa8774ed8f640e1f582456e0b8f63c2bcc2ceec2d0"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.922928 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.955761 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"96624283ee8ce2650eaa77e705919f89ac9d9575e0ee0417b2d910ff0bd7463e"} Nov 28 11:24:31 crc kubenswrapper[4862]: I1128 11:24:31.955812 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1f43f98b6f55c6afca5dd0a3e37e68cd6672052c01be3034ddfea40ef36049b5"} Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.050082 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.050056 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405475-qkjm5" event={"ID":"3a1a2178-787b-4af4-a15f-225ec25b4216","Type":"ContainerDied","Data":"ee937bcba2a2835eaa2fd43dc9bc9f65a9158b80a528d396d5291454cbefb273"} Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.050834 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee937bcba2a2835eaa2fd43dc9bc9f65a9158b80a528d396d5291454cbefb273" Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.061611 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e8740934-8747-4cea-85f9-8e127089df8e","Type":"ContainerStarted","Data":"3856ce9ad332f39780d08b59f8e324869a5f7456fc4aa8f9a8214886c2e2d2cc"} Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.073217 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cld25" Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.469130 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.540138 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kubelet-dir\") pod \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\" (UID: \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\") " Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.540410 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8b7604a9-e2aa-4d7b-b467-0d38d992ab93" (UID: "8b7604a9-e2aa-4d7b-b467-0d38d992ab93"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.540524 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kube-api-access\") pod \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\" (UID: \"8b7604a9-e2aa-4d7b-b467-0d38d992ab93\") " Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.540977 4862 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.554178 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8b7604a9-e2aa-4d7b-b467-0d38d992ab93" (UID: "8b7604a9-e2aa-4d7b-b467-0d38d992ab93"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.644256 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b7604a9-e2aa-4d7b-b467-0d38d992ab93-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.803550 4862 patch_prober.go:28] interesting pod/router-default-5444994796-2bhn7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 11:24:32 crc kubenswrapper[4862]: [-]has-synced failed: reason withheld Nov 28 11:24:32 crc kubenswrapper[4862]: [+]process-running ok Nov 28 11:24:32 crc kubenswrapper[4862]: healthz check failed Nov 28 11:24:32 crc kubenswrapper[4862]: I1128 11:24:32.803623 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2bhn7" podUID="269efc51-3dfb-4888-bc8f-4f681ac5d2c6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.093339 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e8740934-8747-4cea-85f9-8e127089df8e","Type":"ContainerStarted","Data":"608151ed58437ab0664b7dc72a622045112e59d1b6a2d94c1299f0916898b9f3"} Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.107832 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2nbn" event={"ID":"e9947d52-a469-4ca3-b6b5-9482bf691c4c","Type":"ContainerDied","Data":"5df8f6e3fbb3eab241eba05fb3fabe95469a8d4aca35bd26faf60007749be510"} Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.108363 4862 generic.go:334] "Generic (PLEG): container finished" podID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerID="5df8f6e3fbb3eab241eba05fb3fabe95469a8d4aca35bd26faf60007749be510" exitCode=0 Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.131566 4862 generic.go:334] "Generic (PLEG): container finished" podID="02589a29-1121-48de-98a6-691d4f8c598f" containerID="08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480" exitCode=0 Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.132041 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w992j" event={"ID":"02589a29-1121-48de-98a6-691d4f8c598f","Type":"ContainerDied","Data":"08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480"} Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.138744 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.139182 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8b7604a9-e2aa-4d7b-b467-0d38d992ab93","Type":"ContainerDied","Data":"e137a3c738993d7a76bf1dc99336b1f4936a51092df2791b741fee11a6a6e9a8"} Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.139425 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e137a3c738993d7a76bf1dc99336b1f4936a51092df2791b741fee11a6a6e9a8" Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.800725 4862 patch_prober.go:28] interesting pod/router-default-5444994796-2bhn7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 11:24:33 crc kubenswrapper[4862]: [-]has-synced failed: reason withheld Nov 28 11:24:33 crc kubenswrapper[4862]: [+]process-running ok Nov 28 11:24:33 crc kubenswrapper[4862]: healthz check failed Nov 28 11:24:33 crc kubenswrapper[4862]: I1128 11:24:33.800814 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2bhn7" podUID="269efc51-3dfb-4888-bc8f-4f681ac5d2c6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.168302 4862 generic.go:334] "Generic (PLEG): container finished" podID="e8740934-8747-4cea-85f9-8e127089df8e" containerID="608151ed58437ab0664b7dc72a622045112e59d1b6a2d94c1299f0916898b9f3" exitCode=0 Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.168358 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e8740934-8747-4cea-85f9-8e127089df8e","Type":"ContainerDied","Data":"608151ed58437ab0664b7dc72a622045112e59d1b6a2d94c1299f0916898b9f3"} Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.630388 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.781237 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e8740934-8747-4cea-85f9-8e127089df8e-kubelet-dir\") pod \"e8740934-8747-4cea-85f9-8e127089df8e\" (UID: \"e8740934-8747-4cea-85f9-8e127089df8e\") " Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.781379 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e8740934-8747-4cea-85f9-8e127089df8e-kube-api-access\") pod \"e8740934-8747-4cea-85f9-8e127089df8e\" (UID: \"e8740934-8747-4cea-85f9-8e127089df8e\") " Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.781484 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8740934-8747-4cea-85f9-8e127089df8e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e8740934-8747-4cea-85f9-8e127089df8e" (UID: "e8740934-8747-4cea-85f9-8e127089df8e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.781882 4862 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e8740934-8747-4cea-85f9-8e127089df8e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.793643 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8740934-8747-4cea-85f9-8e127089df8e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e8740934-8747-4cea-85f9-8e127089df8e" (UID: "e8740934-8747-4cea-85f9-8e127089df8e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.810714 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.816999 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-2bhn7" Nov 28 11:24:34 crc kubenswrapper[4862]: I1128 11:24:34.884055 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e8740934-8747-4cea-85f9-8e127089df8e-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 11:24:35 crc kubenswrapper[4862]: I1128 11:24:35.207325 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 11:24:35 crc kubenswrapper[4862]: I1128 11:24:35.207715 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e8740934-8747-4cea-85f9-8e127089df8e","Type":"ContainerDied","Data":"3856ce9ad332f39780d08b59f8e324869a5f7456fc4aa8f9a8214886c2e2d2cc"} Nov 28 11:24:35 crc kubenswrapper[4862]: I1128 11:24:35.207739 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3856ce9ad332f39780d08b59f8e324869a5f7456fc4aa8f9a8214886c2e2d2cc" Nov 28 11:24:37 crc kubenswrapper[4862]: I1128 11:24:37.043349 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-kpm6d" Nov 28 11:24:38 crc kubenswrapper[4862]: I1128 11:24:38.305336 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:24:38 crc kubenswrapper[4862]: I1128 11:24:38.305947 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:24:41 crc kubenswrapper[4862]: I1128 11:24:41.107630 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:41 crc kubenswrapper[4862]: I1128 11:24:41.111995 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:24:41 crc kubenswrapper[4862]: I1128 11:24:41.157234 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-45tfc" Nov 28 11:24:44 crc kubenswrapper[4862]: I1128 11:24:44.871688 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:44 crc kubenswrapper[4862]: I1128 11:24:44.880709 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60f0df65-b253-4ea2-abda-dbc50f7f2eca-metrics-certs\") pod \"network-metrics-daemon-wl4cj\" (UID: \"60f0df65-b253-4ea2-abda-dbc50f7f2eca\") " pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:45 crc kubenswrapper[4862]: I1128 11:24:45.094414 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl4cj" Nov 28 11:24:49 crc kubenswrapper[4862]: I1128 11:24:49.448720 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:24:57 crc kubenswrapper[4862]: I1128 11:24:57.185605 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-wl4cj"] Nov 28 11:25:01 crc kubenswrapper[4862]: I1128 11:25:01.319595 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wndxn" Nov 28 11:25:02 crc kubenswrapper[4862]: E1128 11:25:02.040611 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 28 11:25:02 crc kubenswrapper[4862]: E1128 11:25:02.041406 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wfzcg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-wbstd_openshift-marketplace(2f359e91-8745-46c0-acad-6bc9795f04ba): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 11:25:02 crc kubenswrapper[4862]: E1128 11:25:02.042835 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-wbstd" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" Nov 28 11:25:03 crc kubenswrapper[4862]: E1128 11:25:03.191713 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 11:25:03 crc kubenswrapper[4862]: E1128 11:25:03.192288 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ppjfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2rcg7_openshift-marketplace(c4b55241-4e78-4d4b-a1ea-55537f8c8ff4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 11:25:03 crc kubenswrapper[4862]: E1128 11:25:03.193503 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2rcg7" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.317797 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 11:25:03 crc kubenswrapper[4862]: E1128 11:25:03.318257 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8740934-8747-4cea-85f9-8e127089df8e" containerName="pruner" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.318274 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8740934-8747-4cea-85f9-8e127089df8e" containerName="pruner" Nov 28 11:25:03 crc kubenswrapper[4862]: E1128 11:25:03.318287 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a1a2178-787b-4af4-a15f-225ec25b4216" containerName="collect-profiles" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.318293 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a1a2178-787b-4af4-a15f-225ec25b4216" containerName="collect-profiles" Nov 28 11:25:03 crc kubenswrapper[4862]: E1128 11:25:03.318304 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b7604a9-e2aa-4d7b-b467-0d38d992ab93" containerName="pruner" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.318310 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b7604a9-e2aa-4d7b-b467-0d38d992ab93" containerName="pruner" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.318415 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a1a2178-787b-4af4-a15f-225ec25b4216" containerName="collect-profiles" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.318427 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b7604a9-e2aa-4d7b-b467-0d38d992ab93" containerName="pruner" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.318437 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8740934-8747-4cea-85f9-8e127089df8e" containerName="pruner" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.318897 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.329661 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.329797 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.336955 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.473308 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37f03ec6-c596-43b7-8f37-dabb96df0858-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"37f03ec6-c596-43b7-8f37-dabb96df0858\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.473473 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37f03ec6-c596-43b7-8f37-dabb96df0858-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"37f03ec6-c596-43b7-8f37-dabb96df0858\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.574973 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37f03ec6-c596-43b7-8f37-dabb96df0858-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"37f03ec6-c596-43b7-8f37-dabb96df0858\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.575122 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37f03ec6-c596-43b7-8f37-dabb96df0858-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"37f03ec6-c596-43b7-8f37-dabb96df0858\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.575515 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37f03ec6-c596-43b7-8f37-dabb96df0858-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"37f03ec6-c596-43b7-8f37-dabb96df0858\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.612065 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37f03ec6-c596-43b7-8f37-dabb96df0858-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"37f03ec6-c596-43b7-8f37-dabb96df0858\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:03 crc kubenswrapper[4862]: I1128 11:25:03.654736 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.670273 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.671000 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tz9jc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4hzfd_openshift-marketplace(6edecb40-afd9-40ab-8f5b-227730410e66): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.672148 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4hzfd" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.674060 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.674196 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m7cmt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-l8jkt_openshift-marketplace(9ded77a9-e0a3-410f-b118-8ab6e7be7bcf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.675370 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-l8jkt" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.713200 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.713368 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kwbpj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-w992j_openshift-marketplace(02589a29-1121-48de-98a6-691d4f8c598f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.714529 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-w992j" podUID="02589a29-1121-48de-98a6-691d4f8c598f" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.773333 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.773528 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6nszl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-c2nbn_openshift-marketplace(e9947d52-a469-4ca3-b6b5-9482bf691c4c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 11:25:06 crc kubenswrapper[4862]: E1128 11:25:06.774686 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-c2nbn" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" Nov 28 11:25:07 crc kubenswrapper[4862]: E1128 11:25:07.911549 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2rcg7" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" Nov 28 11:25:07 crc kubenswrapper[4862]: E1128 11:25:07.911618 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-wbstd" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" Nov 28 11:25:07 crc kubenswrapper[4862]: E1128 11:25:07.912354 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-c2nbn" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" Nov 28 11:25:07 crc kubenswrapper[4862]: E1128 11:25:07.912374 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-w992j" podUID="02589a29-1121-48de-98a6-691d4f8c598f" Nov 28 11:25:07 crc kubenswrapper[4862]: E1128 11:25:07.912525 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4hzfd" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" Nov 28 11:25:07 crc kubenswrapper[4862]: E1128 11:25:07.912691 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-l8jkt" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" Nov 28 11:25:07 crc kubenswrapper[4862]: W1128 11:25:07.913740 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60f0df65_b253_4ea2_abda_dbc50f7f2eca.slice/crio-163b9c4b0f501b1f19e04fea15e7730441fd93707b5ac5583133721583cf049a WatchSource:0}: Error finding container 163b9c4b0f501b1f19e04fea15e7730441fd93707b5ac5583133721583cf049a: Status 404 returned error can't find the container with id 163b9c4b0f501b1f19e04fea15e7730441fd93707b5ac5583133721583cf049a Nov 28 11:25:08 crc kubenswrapper[4862]: E1128 11:25:07.997740 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 11:25:08 crc kubenswrapper[4862]: E1128 11:25:07.997947 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-djjls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-cph87_openshift-marketplace(22c7db45-0f31-4bc7-b93d-4470c51434b1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 11:25:08 crc kubenswrapper[4862]: E1128 11:25:07.998735 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 11:25:08 crc kubenswrapper[4862]: E1128 11:25:07.998929 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tvgdr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-j5c55_openshift-marketplace(19d8710e-5f52-452e-8b13-854546736587): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 11:25:08 crc kubenswrapper[4862]: E1128 11:25:07.999342 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-cph87" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" Nov 28 11:25:08 crc kubenswrapper[4862]: E1128 11:25:08.000661 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-j5c55" podUID="19d8710e-5f52-452e-8b13-854546736587" Nov 28 11:25:08 crc kubenswrapper[4862]: I1128 11:25:08.112931 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 11:25:08 crc kubenswrapper[4862]: W1128 11:25:08.133558 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod37f03ec6_c596_43b7_8f37_dabb96df0858.slice/crio-e31d3d6bd8f591fbf8c64bc01db09ab5aa8bdc5c8676ac4211a03bb11dc3fb26 WatchSource:0}: Error finding container e31d3d6bd8f591fbf8c64bc01db09ab5aa8bdc5c8676ac4211a03bb11dc3fb26: Status 404 returned error can't find the container with id e31d3d6bd8f591fbf8c64bc01db09ab5aa8bdc5c8676ac4211a03bb11dc3fb26 Nov 28 11:25:08 crc kubenswrapper[4862]: I1128 11:25:08.291859 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:25:08 crc kubenswrapper[4862]: I1128 11:25:08.292389 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:25:08 crc kubenswrapper[4862]: I1128 11:25:08.470496 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" event={"ID":"60f0df65-b253-4ea2-abda-dbc50f7f2eca","Type":"ContainerStarted","Data":"618864f79729fc8fe763d921fddf8019c404dc84973e3108c0fbc9bfd63c740b"} Nov 28 11:25:08 crc kubenswrapper[4862]: I1128 11:25:08.470952 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" event={"ID":"60f0df65-b253-4ea2-abda-dbc50f7f2eca","Type":"ContainerStarted","Data":"163b9c4b0f501b1f19e04fea15e7730441fd93707b5ac5583133721583cf049a"} Nov 28 11:25:08 crc kubenswrapper[4862]: I1128 11:25:08.472069 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"37f03ec6-c596-43b7-8f37-dabb96df0858","Type":"ContainerStarted","Data":"e31d3d6bd8f591fbf8c64bc01db09ab5aa8bdc5c8676ac4211a03bb11dc3fb26"} Nov 28 11:25:08 crc kubenswrapper[4862]: E1128 11:25:08.473923 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-j5c55" podUID="19d8710e-5f52-452e-8b13-854546736587" Nov 28 11:25:08 crc kubenswrapper[4862]: E1128 11:25:08.476852 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-cph87" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.480678 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-wl4cj" event={"ID":"60f0df65-b253-4ea2-abda-dbc50f7f2eca","Type":"ContainerStarted","Data":"a7b515cd7e88ab209e048fc387a8218a7657988051b14f028b3745a2e7a938d2"} Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.482735 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"37f03ec6-c596-43b7-8f37-dabb96df0858","Type":"ContainerStarted","Data":"6eaf63f652e433d774647db7a24f4de1bde0977245766feefb080e914f1bf852"} Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.502697 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-wl4cj" podStartSLOduration=167.502669587 podStartE2EDuration="2m47.502669587s" podCreationTimestamp="2025-11-28 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:25:09.498194585 +0000 UTC m=+187.030708536" watchObservedRunningTime="2025-11-28 11:25:09.502669587 +0000 UTC m=+187.035183498" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.518620 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.519494 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.532315 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=6.532290779 podStartE2EDuration="6.532290779s" podCreationTimestamp="2025-11-28 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:25:09.531697111 +0000 UTC m=+187.064211052" watchObservedRunningTime="2025-11-28 11:25:09.532290779 +0000 UTC m=+187.064804710" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.546457 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.572697 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-var-lock\") pod \"installer-9-crc\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.572785 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b644bbd7-4723-4b57-9844-d956bbbb964c-kube-api-access\") pod \"installer-9-crc\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.573032 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.674830 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b644bbd7-4723-4b57-9844-d956bbbb964c-kube-api-access\") pod \"installer-9-crc\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.674974 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.675052 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-var-lock\") pod \"installer-9-crc\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.675178 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-var-lock\") pod \"installer-9-crc\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.675216 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.702643 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b644bbd7-4723-4b57-9844-d956bbbb964c-kube-api-access\") pod \"installer-9-crc\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:09 crc kubenswrapper[4862]: I1128 11:25:09.841745 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:10 crc kubenswrapper[4862]: I1128 11:25:10.108433 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 11:25:10 crc kubenswrapper[4862]: I1128 11:25:10.117462 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 11:25:10 crc kubenswrapper[4862]: I1128 11:25:10.490394 4862 generic.go:334] "Generic (PLEG): container finished" podID="37f03ec6-c596-43b7-8f37-dabb96df0858" containerID="6eaf63f652e433d774647db7a24f4de1bde0977245766feefb080e914f1bf852" exitCode=0 Nov 28 11:25:10 crc kubenswrapper[4862]: I1128 11:25:10.490544 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"37f03ec6-c596-43b7-8f37-dabb96df0858","Type":"ContainerDied","Data":"6eaf63f652e433d774647db7a24f4de1bde0977245766feefb080e914f1bf852"} Nov 28 11:25:10 crc kubenswrapper[4862]: I1128 11:25:10.494660 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b644bbd7-4723-4b57-9844-d956bbbb964c","Type":"ContainerStarted","Data":"645c4e02370aee111b05bed7512b74190e5dcd59095b935f7f4911dc63b00ac0"} Nov 28 11:25:10 crc kubenswrapper[4862]: I1128 11:25:10.494694 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b644bbd7-4723-4b57-9844-d956bbbb964c","Type":"ContainerStarted","Data":"5c24e67443ede01286322b672e865e796e6880422cbaa344846e062c51df3ae2"} Nov 28 11:25:10 crc kubenswrapper[4862]: I1128 11:25:10.528514 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.528494332 podStartE2EDuration="1.528494332s" podCreationTimestamp="2025-11-28 11:25:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:25:10.524313339 +0000 UTC m=+188.056827260" watchObservedRunningTime="2025-11-28 11:25:10.528494332 +0000 UTC m=+188.061008243" Nov 28 11:25:11 crc kubenswrapper[4862]: I1128 11:25:11.778845 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:11 crc kubenswrapper[4862]: I1128 11:25:11.911340 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37f03ec6-c596-43b7-8f37-dabb96df0858-kube-api-access\") pod \"37f03ec6-c596-43b7-8f37-dabb96df0858\" (UID: \"37f03ec6-c596-43b7-8f37-dabb96df0858\") " Nov 28 11:25:11 crc kubenswrapper[4862]: I1128 11:25:11.911467 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37f03ec6-c596-43b7-8f37-dabb96df0858-kubelet-dir\") pod \"37f03ec6-c596-43b7-8f37-dabb96df0858\" (UID: \"37f03ec6-c596-43b7-8f37-dabb96df0858\") " Nov 28 11:25:11 crc kubenswrapper[4862]: I1128 11:25:11.913523 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37f03ec6-c596-43b7-8f37-dabb96df0858-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "37f03ec6-c596-43b7-8f37-dabb96df0858" (UID: "37f03ec6-c596-43b7-8f37-dabb96df0858"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:25:11 crc kubenswrapper[4862]: I1128 11:25:11.920214 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37f03ec6-c596-43b7-8f37-dabb96df0858-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "37f03ec6-c596-43b7-8f37-dabb96df0858" (UID: "37f03ec6-c596-43b7-8f37-dabb96df0858"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:25:12 crc kubenswrapper[4862]: I1128 11:25:12.013730 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37f03ec6-c596-43b7-8f37-dabb96df0858-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:12 crc kubenswrapper[4862]: I1128 11:25:12.013771 4862 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37f03ec6-c596-43b7-8f37-dabb96df0858-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:12 crc kubenswrapper[4862]: I1128 11:25:12.508455 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"37f03ec6-c596-43b7-8f37-dabb96df0858","Type":"ContainerDied","Data":"e31d3d6bd8f591fbf8c64bc01db09ab5aa8bdc5c8676ac4211a03bb11dc3fb26"} Nov 28 11:25:12 crc kubenswrapper[4862]: I1128 11:25:12.508496 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e31d3d6bd8f591fbf8c64bc01db09ab5aa8bdc5c8676ac4211a03bb11dc3fb26" Nov 28 11:25:12 crc kubenswrapper[4862]: I1128 11:25:12.508581 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 11:25:21 crc kubenswrapper[4862]: I1128 11:25:21.653336 4862 generic.go:334] "Generic (PLEG): container finished" podID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerID="3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a" exitCode=0 Nov 28 11:25:21 crc kubenswrapper[4862]: I1128 11:25:21.653429 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8jkt" event={"ID":"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf","Type":"ContainerDied","Data":"3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a"} Nov 28 11:25:21 crc kubenswrapper[4862]: I1128 11:25:21.659400 4862 generic.go:334] "Generic (PLEG): container finished" podID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerID="7d5de1a7d6f180c4f4a8017232661637cca7b974bf6381e233ef101fcb3ce652" exitCode=0 Nov 28 11:25:21 crc kubenswrapper[4862]: I1128 11:25:21.659512 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2nbn" event={"ID":"e9947d52-a469-4ca3-b6b5-9482bf691c4c","Type":"ContainerDied","Data":"7d5de1a7d6f180c4f4a8017232661637cca7b974bf6381e233ef101fcb3ce652"} Nov 28 11:25:21 crc kubenswrapper[4862]: I1128 11:25:21.662255 4862 generic.go:334] "Generic (PLEG): container finished" podID="02589a29-1121-48de-98a6-691d4f8c598f" containerID="34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446" exitCode=0 Nov 28 11:25:21 crc kubenswrapper[4862]: I1128 11:25:21.662293 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w992j" event={"ID":"02589a29-1121-48de-98a6-691d4f8c598f","Type":"ContainerDied","Data":"34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446"} Nov 28 11:25:21 crc kubenswrapper[4862]: I1128 11:25:21.664902 4862 generic.go:334] "Generic (PLEG): container finished" podID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerID="faa354060b696b1d143863519bd427d0390e78410605b8f4749c3d0780ffdf60" exitCode=0 Nov 28 11:25:21 crc kubenswrapper[4862]: I1128 11:25:21.667077 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbstd" event={"ID":"2f359e91-8745-46c0-acad-6bc9795f04ba","Type":"ContainerDied","Data":"faa354060b696b1d143863519bd427d0390e78410605b8f4749c3d0780ffdf60"} Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.673532 4862 generic.go:334] "Generic (PLEG): container finished" podID="6edecb40-afd9-40ab-8f5b-227730410e66" containerID="4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d" exitCode=0 Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.673623 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hzfd" event={"ID":"6edecb40-afd9-40ab-8f5b-227730410e66","Type":"ContainerDied","Data":"4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d"} Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.677482 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbstd" event={"ID":"2f359e91-8745-46c0-acad-6bc9795f04ba","Type":"ContainerStarted","Data":"978c2dcc8a8182f2b01e80eeed8d829ad183f70dc89b51af05b1b3f05a6a4468"} Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.680306 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8jkt" event={"ID":"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf","Type":"ContainerStarted","Data":"059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714"} Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.682881 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2nbn" event={"ID":"e9947d52-a469-4ca3-b6b5-9482bf691c4c","Type":"ContainerStarted","Data":"3374373ace813cb6f4a57249b55f5b4cda37a328052b0ce3cbc3dd054897fc5a"} Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.685334 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w992j" event={"ID":"02589a29-1121-48de-98a6-691d4f8c598f","Type":"ContainerStarted","Data":"a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0"} Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.710304 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l8jkt" podStartSLOduration=2.282942351 podStartE2EDuration="53.710287477s" podCreationTimestamp="2025-11-28 11:24:29 +0000 UTC" firstStartedPulling="2025-11-28 11:24:30.813278753 +0000 UTC m=+148.345792674" lastFinishedPulling="2025-11-28 11:25:22.240623869 +0000 UTC m=+199.773137800" observedRunningTime="2025-11-28 11:25:22.709743151 +0000 UTC m=+200.242257092" watchObservedRunningTime="2025-11-28 11:25:22.710287477 +0000 UTC m=+200.242801388" Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.733875 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wbstd" podStartSLOduration=3.141365385 podStartE2EDuration="56.733860281s" podCreationTimestamp="2025-11-28 11:24:26 +0000 UTC" firstStartedPulling="2025-11-28 11:24:28.56278241 +0000 UTC m=+146.095296331" lastFinishedPulling="2025-11-28 11:25:22.155277306 +0000 UTC m=+199.687791227" observedRunningTime="2025-11-28 11:25:22.731918774 +0000 UTC m=+200.264432695" watchObservedRunningTime="2025-11-28 11:25:22.733860281 +0000 UTC m=+200.266374202" Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.774053 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w992j" podStartSLOduration=3.462804175 podStartE2EDuration="53.774036364s" podCreationTimestamp="2025-11-28 11:24:29 +0000 UTC" firstStartedPulling="2025-11-28 11:24:31.904206809 +0000 UTC m=+149.436720730" lastFinishedPulling="2025-11-28 11:25:22.215438998 +0000 UTC m=+199.747952919" observedRunningTime="2025-11-28 11:25:22.771707025 +0000 UTC m=+200.304220946" watchObservedRunningTime="2025-11-28 11:25:22.774036364 +0000 UTC m=+200.306550285" Nov 28 11:25:22 crc kubenswrapper[4862]: I1128 11:25:22.774793 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c2nbn" podStartSLOduration=2.507121669 podStartE2EDuration="52.774788786s" podCreationTimestamp="2025-11-28 11:24:30 +0000 UTC" firstStartedPulling="2025-11-28 11:24:31.873945417 +0000 UTC m=+149.406459338" lastFinishedPulling="2025-11-28 11:25:22.141612534 +0000 UTC m=+199.674126455" observedRunningTime="2025-11-28 11:25:22.753489049 +0000 UTC m=+200.286002970" watchObservedRunningTime="2025-11-28 11:25:22.774788786 +0000 UTC m=+200.307302707" Nov 28 11:25:27 crc kubenswrapper[4862]: I1128 11:25:27.067950 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:25:27 crc kubenswrapper[4862]: I1128 11:25:27.068561 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:25:27 crc kubenswrapper[4862]: I1128 11:25:27.351813 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:25:27 crc kubenswrapper[4862]: I1128 11:25:27.763987 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:25:29 crc kubenswrapper[4862]: I1128 11:25:29.448257 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:25:29 crc kubenswrapper[4862]: I1128 11:25:29.448541 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:25:29 crc kubenswrapper[4862]: I1128 11:25:29.517304 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:25:29 crc kubenswrapper[4862]: I1128 11:25:29.773417 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:25:30 crc kubenswrapper[4862]: I1128 11:25:30.312818 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g5fhw"] Nov 28 11:25:30 crc kubenswrapper[4862]: I1128 11:25:30.353430 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:25:30 crc kubenswrapper[4862]: I1128 11:25:30.354962 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:25:30 crc kubenswrapper[4862]: I1128 11:25:30.420390 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:25:30 crc kubenswrapper[4862]: I1128 11:25:30.443355 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:25:30 crc kubenswrapper[4862]: I1128 11:25:30.443432 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:25:30 crc kubenswrapper[4862]: I1128 11:25:30.502237 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:25:30 crc kubenswrapper[4862]: I1128 11:25:30.781052 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:25:30 crc kubenswrapper[4862]: I1128 11:25:30.803967 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.082972 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c2nbn"] Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.083956 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c2nbn" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerName="registry-server" containerID="cri-o://3374373ace813cb6f4a57249b55f5b4cda37a328052b0ce3cbc3dd054897fc5a" gracePeriod=2 Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.789654 4862 generic.go:334] "Generic (PLEG): container finished" podID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerID="8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7" exitCode=0 Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.789763 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2rcg7" event={"ID":"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4","Type":"ContainerDied","Data":"8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7"} Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.792306 4862 generic.go:334] "Generic (PLEG): container finished" podID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerID="31b511f511059bf076368b01a7cf350621fef7692bcc48216e0afd850541b00f" exitCode=0 Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.792391 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cph87" event={"ID":"22c7db45-0f31-4bc7-b93d-4470c51434b1","Type":"ContainerDied","Data":"31b511f511059bf076368b01a7cf350621fef7692bcc48216e0afd850541b00f"} Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.835776 4862 generic.go:334] "Generic (PLEG): container finished" podID="19d8710e-5f52-452e-8b13-854546736587" containerID="ac8f2909e2d2f83d2a9587c9c051fdc44a2111f5ba3ce35134ef8ae5d1c19bb0" exitCode=0 Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.835896 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5c55" event={"ID":"19d8710e-5f52-452e-8b13-854546736587","Type":"ContainerDied","Data":"ac8f2909e2d2f83d2a9587c9c051fdc44a2111f5ba3ce35134ef8ae5d1c19bb0"} Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.843168 4862 generic.go:334] "Generic (PLEG): container finished" podID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerID="3374373ace813cb6f4a57249b55f5b4cda37a328052b0ce3cbc3dd054897fc5a" exitCode=0 Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.843285 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2nbn" event={"ID":"e9947d52-a469-4ca3-b6b5-9482bf691c4c","Type":"ContainerDied","Data":"3374373ace813cb6f4a57249b55f5b4cda37a328052b0ce3cbc3dd054897fc5a"} Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.853726 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hzfd" event={"ID":"6edecb40-afd9-40ab-8f5b-227730410e66","Type":"ContainerStarted","Data":"70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c"} Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.909748 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4hzfd" podStartSLOduration=3.960903027 podStartE2EDuration="1m6.909721343s" podCreationTimestamp="2025-11-28 11:24:27 +0000 UTC" firstStartedPulling="2025-11-28 11:24:29.63091111 +0000 UTC m=+147.163425031" lastFinishedPulling="2025-11-28 11:25:32.579729386 +0000 UTC m=+210.112243347" observedRunningTime="2025-11-28 11:25:33.904252342 +0000 UTC m=+211.436766303" watchObservedRunningTime="2025-11-28 11:25:33.909721343 +0000 UTC m=+211.442235294" Nov 28 11:25:33 crc kubenswrapper[4862]: I1128 11:25:33.988198 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.065391 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-catalog-content\") pod \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.065581 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nszl\" (UniqueName: \"kubernetes.io/projected/e9947d52-a469-4ca3-b6b5-9482bf691c4c-kube-api-access-6nszl\") pod \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.075917 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9947d52-a469-4ca3-b6b5-9482bf691c4c-kube-api-access-6nszl" (OuterVolumeSpecName: "kube-api-access-6nszl") pod "e9947d52-a469-4ca3-b6b5-9482bf691c4c" (UID: "e9947d52-a469-4ca3-b6b5-9482bf691c4c"). InnerVolumeSpecName "kube-api-access-6nszl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.167492 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-utilities\") pod \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\" (UID: \"e9947d52-a469-4ca3-b6b5-9482bf691c4c\") " Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.167951 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nszl\" (UniqueName: \"kubernetes.io/projected/e9947d52-a469-4ca3-b6b5-9482bf691c4c-kube-api-access-6nszl\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.168469 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-utilities" (OuterVolumeSpecName: "utilities") pod "e9947d52-a469-4ca3-b6b5-9482bf691c4c" (UID: "e9947d52-a469-4ca3-b6b5-9482bf691c4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.234983 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9947d52-a469-4ca3-b6b5-9482bf691c4c" (UID: "e9947d52-a469-4ca3-b6b5-9482bf691c4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.270136 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.270414 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9947d52-a469-4ca3-b6b5-9482bf691c4c-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.862834 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2nbn" event={"ID":"e9947d52-a469-4ca3-b6b5-9482bf691c4c","Type":"ContainerDied","Data":"9fb91a6d48b32698e42d33e5175c898a18eb9bb5859f63366b87c744963758f0"} Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.862865 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2nbn" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.862930 4862 scope.go:117] "RemoveContainer" containerID="3374373ace813cb6f4a57249b55f5b4cda37a328052b0ce3cbc3dd054897fc5a" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.867482 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2rcg7" event={"ID":"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4","Type":"ContainerStarted","Data":"e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4"} Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.873506 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cph87" event={"ID":"22c7db45-0f31-4bc7-b93d-4470c51434b1","Type":"ContainerStarted","Data":"15cdd6a93a5ae1894530992e43a38d30e304f4771e10c25cb03ae9617b523d5f"} Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.882295 4862 scope.go:117] "RemoveContainer" containerID="7d5de1a7d6f180c4f4a8017232661637cca7b974bf6381e233ef101fcb3ce652" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.900401 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2rcg7" podStartSLOduration=3.524207644 podStartE2EDuration="1m5.900374479s" podCreationTimestamp="2025-11-28 11:24:29 +0000 UTC" firstStartedPulling="2025-11-28 11:24:31.913308748 +0000 UTC m=+149.445822669" lastFinishedPulling="2025-11-28 11:25:34.289475573 +0000 UTC m=+211.821989504" observedRunningTime="2025-11-28 11:25:34.898481423 +0000 UTC m=+212.430995394" watchObservedRunningTime="2025-11-28 11:25:34.900374479 +0000 UTC m=+212.432888400" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.924460 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c2nbn"] Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.925578 4862 scope.go:117] "RemoveContainer" containerID="5df8f6e3fbb3eab241eba05fb3fabe95469a8d4aca35bd26faf60007749be510" Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.929209 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c2nbn"] Nov 28 11:25:34 crc kubenswrapper[4862]: I1128 11:25:34.944853 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cph87" podStartSLOduration=3.309983512 podStartE2EDuration="1m7.944833668s" podCreationTimestamp="2025-11-28 11:24:27 +0000 UTC" firstStartedPulling="2025-11-28 11:24:29.613546237 +0000 UTC m=+147.146060158" lastFinishedPulling="2025-11-28 11:25:34.248396383 +0000 UTC m=+211.780910314" observedRunningTime="2025-11-28 11:25:34.942938792 +0000 UTC m=+212.475452713" watchObservedRunningTime="2025-11-28 11:25:34.944833668 +0000 UTC m=+212.477347589" Nov 28 11:25:35 crc kubenswrapper[4862]: I1128 11:25:35.897117 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5c55" event={"ID":"19d8710e-5f52-452e-8b13-854546736587","Type":"ContainerStarted","Data":"95abc85c195da8dad6596f05b3370a934f47a49cef7c6bdce53fa1ddacab0611"} Nov 28 11:25:35 crc kubenswrapper[4862]: I1128 11:25:35.921754 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j5c55" podStartSLOduration=3.964846438 podStartE2EDuration="1m9.921737559s" podCreationTimestamp="2025-11-28 11:24:26 +0000 UTC" firstStartedPulling="2025-11-28 11:24:28.570730333 +0000 UTC m=+146.103244254" lastFinishedPulling="2025-11-28 11:25:34.527621444 +0000 UTC m=+212.060135375" observedRunningTime="2025-11-28 11:25:35.921264616 +0000 UTC m=+213.453778537" watchObservedRunningTime="2025-11-28 11:25:35.921737559 +0000 UTC m=+213.454251480" Nov 28 11:25:36 crc kubenswrapper[4862]: I1128 11:25:36.852667 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" path="/var/lib/kubelet/pods/e9947d52-a469-4ca3-b6b5-9482bf691c4c/volumes" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.265409 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.265474 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.313440 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.455816 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.455896 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.525969 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.690172 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.691195 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.758673 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:25:37 crc kubenswrapper[4862]: I1128 11:25:37.981612 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:25:38 crc kubenswrapper[4862]: I1128 11:25:38.292258 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:25:38 crc kubenswrapper[4862]: I1128 11:25:38.292650 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:25:38 crc kubenswrapper[4862]: I1128 11:25:38.292717 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:25:38 crc kubenswrapper[4862]: I1128 11:25:38.293669 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141"} pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:25:38 crc kubenswrapper[4862]: I1128 11:25:38.293840 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" containerID="cri-o://02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141" gracePeriod=600 Nov 28 11:25:39 crc kubenswrapper[4862]: I1128 11:25:39.877252 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:25:39 crc kubenswrapper[4862]: I1128 11:25:39.878685 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:25:39 crc kubenswrapper[4862]: I1128 11:25:39.929155 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:25:40 crc kubenswrapper[4862]: I1128 11:25:40.932115 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6845439-1a4c-4f28-9ba4-29149d545995" containerID="02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141" exitCode=0 Nov 28 11:25:40 crc kubenswrapper[4862]: I1128 11:25:40.932135 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerDied","Data":"02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141"} Nov 28 11:25:40 crc kubenswrapper[4862]: I1128 11:25:40.988797 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:25:41 crc kubenswrapper[4862]: I1128 11:25:41.939478 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"b510eb115f59f27520a8dc8f9d8b64d0f26aef3c7c9a1efef94445ccf2f2eb72"} Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.083421 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hzfd"] Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.083675 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4hzfd" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" containerName="registry-server" containerID="cri-o://70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c" gracePeriod=2 Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.431523 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.592685 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz9jc\" (UniqueName: \"kubernetes.io/projected/6edecb40-afd9-40ab-8f5b-227730410e66-kube-api-access-tz9jc\") pod \"6edecb40-afd9-40ab-8f5b-227730410e66\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.592740 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-utilities\") pod \"6edecb40-afd9-40ab-8f5b-227730410e66\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.592855 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-catalog-content\") pod \"6edecb40-afd9-40ab-8f5b-227730410e66\" (UID: \"6edecb40-afd9-40ab-8f5b-227730410e66\") " Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.594384 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-utilities" (OuterVolumeSpecName: "utilities") pod "6edecb40-afd9-40ab-8f5b-227730410e66" (UID: "6edecb40-afd9-40ab-8f5b-227730410e66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.598305 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6edecb40-afd9-40ab-8f5b-227730410e66-kube-api-access-tz9jc" (OuterVolumeSpecName: "kube-api-access-tz9jc") pod "6edecb40-afd9-40ab-8f5b-227730410e66" (UID: "6edecb40-afd9-40ab-8f5b-227730410e66"). InnerVolumeSpecName "kube-api-access-tz9jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.636393 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6edecb40-afd9-40ab-8f5b-227730410e66" (UID: "6edecb40-afd9-40ab-8f5b-227730410e66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.679963 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2rcg7"] Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.694555 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.694608 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz9jc\" (UniqueName: \"kubernetes.io/projected/6edecb40-afd9-40ab-8f5b-227730410e66-kube-api-access-tz9jc\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.694627 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6edecb40-afd9-40ab-8f5b-227730410e66-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.949842 4862 generic.go:334] "Generic (PLEG): container finished" podID="6edecb40-afd9-40ab-8f5b-227730410e66" containerID="70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c" exitCode=0 Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.949953 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hzfd" event={"ID":"6edecb40-afd9-40ab-8f5b-227730410e66","Type":"ContainerDied","Data":"70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c"} Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.949995 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hzfd" Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.950040 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hzfd" event={"ID":"6edecb40-afd9-40ab-8f5b-227730410e66","Type":"ContainerDied","Data":"fa3ebad9e7121fa201a127ee26fc500d31d0db42351e4e9e1d5c1d4ee3a46a04"} Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.950067 4862 scope.go:117] "RemoveContainer" containerID="70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c" Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.950914 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2rcg7" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerName="registry-server" containerID="cri-o://e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4" gracePeriod=2 Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.978519 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hzfd"] Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.983296 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4hzfd"] Nov 28 11:25:42 crc kubenswrapper[4862]: I1128 11:25:42.985661 4862 scope.go:117] "RemoveContainer" containerID="4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.006720 4862 scope.go:117] "RemoveContainer" containerID="808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.112522 4862 scope.go:117] "RemoveContainer" containerID="70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c" Nov 28 11:25:43 crc kubenswrapper[4862]: E1128 11:25:43.114340 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c\": container with ID starting with 70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c not found: ID does not exist" containerID="70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.114417 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c"} err="failed to get container status \"70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c\": rpc error: code = NotFound desc = could not find container \"70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c\": container with ID starting with 70d5252a82b0c38e185e1c42daba155594794f3a18ccb39b95f8536c41ad723c not found: ID does not exist" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.114461 4862 scope.go:117] "RemoveContainer" containerID="4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d" Nov 28 11:25:43 crc kubenswrapper[4862]: E1128 11:25:43.115151 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d\": container with ID starting with 4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d not found: ID does not exist" containerID="4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.115199 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d"} err="failed to get container status \"4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d\": rpc error: code = NotFound desc = could not find container \"4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d\": container with ID starting with 4e51da87cae3ec79c7e0317eef34929d9bbeac6ac872d02cc4db0d409713fb6d not found: ID does not exist" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.115236 4862 scope.go:117] "RemoveContainer" containerID="808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca" Nov 28 11:25:43 crc kubenswrapper[4862]: E1128 11:25:43.116327 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca\": container with ID starting with 808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca not found: ID does not exist" containerID="808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.116373 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca"} err="failed to get container status \"808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca\": rpc error: code = NotFound desc = could not find container \"808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca\": container with ID starting with 808f4d98bd2c68a2afcc06dce7c267118db94b2c76bf89427e20e3e58f1f49ca not found: ID does not exist" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.326577 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.504705 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppjfb\" (UniqueName: \"kubernetes.io/projected/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-kube-api-access-ppjfb\") pod \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.505161 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-utilities\") pod \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.505408 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-catalog-content\") pod \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\" (UID: \"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4\") " Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.506624 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-utilities" (OuterVolumeSpecName: "utilities") pod "c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" (UID: "c4b55241-4e78-4d4b-a1ea-55537f8c8ff4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.511110 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-kube-api-access-ppjfb" (OuterVolumeSpecName: "kube-api-access-ppjfb") pod "c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" (UID: "c4b55241-4e78-4d4b-a1ea-55537f8c8ff4"). InnerVolumeSpecName "kube-api-access-ppjfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.524674 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" (UID: "c4b55241-4e78-4d4b-a1ea-55537f8c8ff4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.606838 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.606884 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.606898 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppjfb\" (UniqueName: \"kubernetes.io/projected/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4-kube-api-access-ppjfb\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.963300 4862 generic.go:334] "Generic (PLEG): container finished" podID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerID="e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4" exitCode=0 Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.963379 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2rcg7" event={"ID":"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4","Type":"ContainerDied","Data":"e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4"} Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.963408 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2rcg7" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.963446 4862 scope.go:117] "RemoveContainer" containerID="e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4" Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.963427 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2rcg7" event={"ID":"c4b55241-4e78-4d4b-a1ea-55537f8c8ff4","Type":"ContainerDied","Data":"5e38ae585a6868cc47c3462c63721b500a0c0717daac8d8cb8fb75d600c8d13f"} Nov 28 11:25:43 crc kubenswrapper[4862]: I1128 11:25:43.982214 4862 scope.go:117] "RemoveContainer" containerID="8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7" Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.008998 4862 scope.go:117] "RemoveContainer" containerID="39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf" Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.017650 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2rcg7"] Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.023867 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2rcg7"] Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.034782 4862 scope.go:117] "RemoveContainer" containerID="e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4" Nov 28 11:25:44 crc kubenswrapper[4862]: E1128 11:25:44.035235 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4\": container with ID starting with e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4 not found: ID does not exist" containerID="e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4" Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.035298 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4"} err="failed to get container status \"e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4\": rpc error: code = NotFound desc = could not find container \"e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4\": container with ID starting with e1dc405f9b0a87f66c20d5a89b9746c361cca6e646f3abe9bf61137a78f74cf4 not found: ID does not exist" Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.035417 4862 scope.go:117] "RemoveContainer" containerID="8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7" Nov 28 11:25:44 crc kubenswrapper[4862]: E1128 11:25:44.035804 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7\": container with ID starting with 8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7 not found: ID does not exist" containerID="8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7" Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.035841 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7"} err="failed to get container status \"8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7\": rpc error: code = NotFound desc = could not find container \"8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7\": container with ID starting with 8390ba6b28b5294338d7f9839ea4d137377a38819dfb7d0ed904b54876123be7 not found: ID does not exist" Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.035881 4862 scope.go:117] "RemoveContainer" containerID="39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf" Nov 28 11:25:44 crc kubenswrapper[4862]: E1128 11:25:44.036570 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf\": container with ID starting with 39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf not found: ID does not exist" containerID="39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf" Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.036622 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf"} err="failed to get container status \"39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf\": rpc error: code = NotFound desc = could not find container \"39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf\": container with ID starting with 39fe8329560b1bb7b2bc6d8ba249ae5c89ccac77f063f0379db5240fe90dbacf not found: ID does not exist" Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.851895 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" path="/var/lib/kubelet/pods/6edecb40-afd9-40ab-8f5b-227730410e66/volumes" Nov 28 11:25:44 crc kubenswrapper[4862]: I1128 11:25:44.853966 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" path="/var/lib/kubelet/pods/c4b55241-4e78-4d4b-a1ea-55537f8c8ff4/volumes" Nov 28 11:25:47 crc kubenswrapper[4862]: I1128 11:25:47.338558 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:25:47 crc kubenswrapper[4862]: I1128 11:25:47.743862 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.452608 4862 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.452900 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerName="registry-server" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.452914 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerName="registry-server" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.452930 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" containerName="extract-content" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.452938 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" containerName="extract-content" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.452955 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37f03ec6-c596-43b7-8f37-dabb96df0858" containerName="pruner" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.452963 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="37f03ec6-c596-43b7-8f37-dabb96df0858" containerName="pruner" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.452974 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerName="extract-utilities" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.452984 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerName="extract-utilities" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.452995 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerName="extract-utilities" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453003 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerName="extract-utilities" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.453017 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" containerName="extract-utilities" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453025 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" containerName="extract-utilities" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.453035 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerName="extract-content" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453043 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerName="extract-content" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.453053 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" containerName="registry-server" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453062 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" containerName="registry-server" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.453075 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerName="registry-server" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453083 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerName="registry-server" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.453117 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerName="extract-content" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453128 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerName="extract-content" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453265 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9947d52-a469-4ca3-b6b5-9482bf691c4c" containerName="registry-server" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453278 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4b55241-4e78-4d4b-a1ea-55537f8c8ff4" containerName="registry-server" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453297 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="37f03ec6-c596-43b7-8f37-dabb96df0858" containerName="pruner" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453307 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="6edecb40-afd9-40ab-8f5b-227730410e66" containerName="registry-server" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453801 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.453923 4862 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.454354 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b" gracePeriod=15 Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.454411 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13" gracePeriod=15 Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.454454 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd" gracePeriod=15 Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.454339 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754" gracePeriod=15 Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.454628 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89" gracePeriod=15 Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.455437 4862 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.455731 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.455745 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.455783 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.455794 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.455809 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.455818 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.455830 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.455839 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.455852 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.455859 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.455871 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.455881 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 11:25:48 crc kubenswrapper[4862]: E1128 11:25:48.455895 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.455904 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.456024 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.456044 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.456055 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.456066 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.456077 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.456104 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.501700 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.588002 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.588066 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.588122 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.588152 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.588174 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.588207 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.588253 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.588277 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689781 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689828 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689858 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689886 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689904 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689922 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689935 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689958 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689959 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689984 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.690000 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689978 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.690012 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.689985 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.690039 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.690026 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: I1128 11:25:48.798127 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:25:48 crc kubenswrapper[4862]: W1128 11:25:48.815233 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-1c360606045d6df13a83c21cbb59859adf7d35613c882bbca7cc7c361a0a0e01 WatchSource:0}: Error finding container 1c360606045d6df13a83c21cbb59859adf7d35613c882bbca7cc7c361a0a0e01: Status 404 returned error can't find the container with id 1c360606045d6df13a83c21cbb59859adf7d35613c882bbca7cc7c361a0a0e01 Nov 28 11:25:49 crc kubenswrapper[4862]: I1128 11:25:49.014427 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"1c360606045d6df13a83c21cbb59859adf7d35613c882bbca7cc7c361a0a0e01"} Nov 28 11:25:49 crc kubenswrapper[4862]: E1128 11:25:49.103810 4862 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187c2803671b5114 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 11:25:49.102248212 +0000 UTC m=+226.634762143,LastTimestamp:2025-11-28 11:25:49.102248212 +0000 UTC m=+226.634762143,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.023732 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565"} Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.024646 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.027013 4862 generic.go:334] "Generic (PLEG): container finished" podID="b644bbd7-4723-4b57-9844-d956bbbb964c" containerID="645c4e02370aee111b05bed7512b74190e5dcd59095b935f7f4911dc63b00ac0" exitCode=0 Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.027133 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b644bbd7-4723-4b57-9844-d956bbbb964c","Type":"ContainerDied","Data":"645c4e02370aee111b05bed7512b74190e5dcd59095b935f7f4911dc63b00ac0"} Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.028690 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.029670 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.030683 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.033130 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.034591 4862 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b" exitCode=0 Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.034664 4862 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13" exitCode=0 Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.034688 4862 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89" exitCode=0 Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.034697 4862 scope.go:117] "RemoveContainer" containerID="a2dc0c0d36ddfb0231de2a5ceebe00dfe95a832036e4cf34c16dc96a97ddba82" Nov 28 11:25:50 crc kubenswrapper[4862]: I1128 11:25:50.034706 4862 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd" exitCode=2 Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.046916 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.398490 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.399511 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.399927 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.549539 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b644bbd7-4723-4b57-9844-d956bbbb964c-kube-api-access\") pod \"b644bbd7-4723-4b57-9844-d956bbbb964c\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.549681 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-var-lock\") pod \"b644bbd7-4723-4b57-9844-d956bbbb964c\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.549819 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-kubelet-dir\") pod \"b644bbd7-4723-4b57-9844-d956bbbb964c\" (UID: \"b644bbd7-4723-4b57-9844-d956bbbb964c\") " Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.549917 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-var-lock" (OuterVolumeSpecName: "var-lock") pod "b644bbd7-4723-4b57-9844-d956bbbb964c" (UID: "b644bbd7-4723-4b57-9844-d956bbbb964c"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.550031 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b644bbd7-4723-4b57-9844-d956bbbb964c" (UID: "b644bbd7-4723-4b57-9844-d956bbbb964c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.550374 4862 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.550413 4862 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b644bbd7-4723-4b57-9844-d956bbbb964c-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.559223 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b644bbd7-4723-4b57-9844-d956bbbb964c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b644bbd7-4723-4b57-9844-d956bbbb964c" (UID: "b644bbd7-4723-4b57-9844-d956bbbb964c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:25:51 crc kubenswrapper[4862]: I1128 11:25:51.651921 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b644bbd7-4723-4b57-9844-d956bbbb964c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.034196 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.035967 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.036416 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.036736 4862 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.037184 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.057576 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.057670 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.057743 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.057932 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.057979 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.058066 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.058772 4862 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.058816 4862 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.058838 4862 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.059544 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b644bbd7-4723-4b57-9844-d956bbbb964c","Type":"ContainerDied","Data":"5c24e67443ede01286322b672e865e796e6880422cbaa344846e062c51df3ae2"} Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.059581 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c24e67443ede01286322b672e865e796e6880422cbaa344846e062c51df3ae2" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.059646 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.066609 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.068779 4862 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754" exitCode=0 Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.068857 4862 scope.go:117] "RemoveContainer" containerID="f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.068905 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.092697 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.092858 4862 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.093008 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.093266 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.093828 4862 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.094215 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.095925 4862 scope.go:117] "RemoveContainer" containerID="eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.112862 4862 scope.go:117] "RemoveContainer" containerID="fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.128978 4862 scope.go:117] "RemoveContainer" containerID="10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.146804 4862 scope.go:117] "RemoveContainer" containerID="baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.171755 4862 scope.go:117] "RemoveContainer" containerID="1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.194452 4862 scope.go:117] "RemoveContainer" containerID="f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b" Nov 28 11:25:52 crc kubenswrapper[4862]: E1128 11:25:52.194922 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\": container with ID starting with f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b not found: ID does not exist" containerID="f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.194986 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b"} err="failed to get container status \"f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\": rpc error: code = NotFound desc = could not find container \"f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b\": container with ID starting with f47cd65d53d0e173f0c3cba62a67989d704871b0eb64c3a55442a679a779112b not found: ID does not exist" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.195014 4862 scope.go:117] "RemoveContainer" containerID="eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13" Nov 28 11:25:52 crc kubenswrapper[4862]: E1128 11:25:52.195587 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\": container with ID starting with eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13 not found: ID does not exist" containerID="eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.196468 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13"} err="failed to get container status \"eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\": rpc error: code = NotFound desc = could not find container \"eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13\": container with ID starting with eec7b226a642e919a9f0b537d5ba4516b0bd94d663f3d27315cc2496c212ce13 not found: ID does not exist" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.196518 4862 scope.go:117] "RemoveContainer" containerID="fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89" Nov 28 11:25:52 crc kubenswrapper[4862]: E1128 11:25:52.197110 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\": container with ID starting with fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89 not found: ID does not exist" containerID="fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.197172 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89"} err="failed to get container status \"fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\": rpc error: code = NotFound desc = could not find container \"fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89\": container with ID starting with fb2868859c94f084ab88f493f656793c50be57226d3c95bb5364c7c2d1852e89 not found: ID does not exist" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.197193 4862 scope.go:117] "RemoveContainer" containerID="10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd" Nov 28 11:25:52 crc kubenswrapper[4862]: E1128 11:25:52.197749 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\": container with ID starting with 10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd not found: ID does not exist" containerID="10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.197798 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd"} err="failed to get container status \"10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\": rpc error: code = NotFound desc = could not find container \"10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd\": container with ID starting with 10f4548b3dc843024d756d9a29ace137d7af97daa4b0a384f5a109d46be706dd not found: ID does not exist" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.197831 4862 scope.go:117] "RemoveContainer" containerID="baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754" Nov 28 11:25:52 crc kubenswrapper[4862]: E1128 11:25:52.198269 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\": container with ID starting with baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754 not found: ID does not exist" containerID="baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.198376 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754"} err="failed to get container status \"baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\": rpc error: code = NotFound desc = could not find container \"baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754\": container with ID starting with baba92ad98836cdb0ce27d220b065e550945ef561cf69a91d2600c3d7ccbf754 not found: ID does not exist" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.198416 4862 scope.go:117] "RemoveContainer" containerID="1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87" Nov 28 11:25:52 crc kubenswrapper[4862]: E1128 11:25:52.198830 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\": container with ID starting with 1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87 not found: ID does not exist" containerID="1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.198864 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87"} err="failed to get container status \"1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\": rpc error: code = NotFound desc = could not find container \"1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87\": container with ID starting with 1838f90e2d9a0654b6696b61c639b74ad13b4d63c37dd21b7da7a1dcfd049a87 not found: ID does not exist" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.841563 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.841898 4862 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.842327 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:52 crc kubenswrapper[4862]: I1128 11:25:52.846737 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 28 11:25:55 crc kubenswrapper[4862]: E1128 11:25:55.173256 4862 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187c2803671b5114 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 11:25:49.102248212 +0000 UTC m=+226.634762143,LastTimestamp:2025-11-28 11:25:49.102248212 +0000 UTC m=+226.634762143,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.364955 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" containerName="oauth-openshift" containerID="cri-o://9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889" gracePeriod=15 Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.794439 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.796018 4862 status_manager.go:851] "Failed to get status for pod" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-g5fhw\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.797253 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.798351 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.815629 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-dir\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.815726 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-service-ca\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.815776 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-router-certs\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.815819 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-policies\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.815854 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-cliconfig\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.815904 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-ocp-branding-template\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.815936 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54gdk\" (UniqueName: \"kubernetes.io/projected/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-kube-api-access-54gdk\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.815980 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-session\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.816032 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-idp-0-file-data\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.816134 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-provider-selection\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.816179 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-error\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.816219 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-trusted-ca-bundle\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.816254 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-login\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.816288 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-serving-cert\") pod \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\" (UID: \"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af\") " Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.817084 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.818642 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.818715 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.820386 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.821516 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.833715 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.835761 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.839736 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.840073 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.840206 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-kube-api-access-54gdk" (OuterVolumeSpecName: "kube-api-access-54gdk") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "kube-api-access-54gdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.840791 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.840848 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.849715 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.849914 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" (UID: "8b34b3ef-6f8f-4785-afd9-aa118a3ef9af"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918156 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918211 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918232 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918251 4862 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918271 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918292 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918310 4862 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918333 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918353 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918374 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54gdk\" (UniqueName: \"kubernetes.io/projected/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-kube-api-access-54gdk\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918393 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918411 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918432 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:55 crc kubenswrapper[4862]: I1128 11:25:55.918453 4862 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.104583 4862 generic.go:334] "Generic (PLEG): container finished" podID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" containerID="9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889" exitCode=0 Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.104629 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" event={"ID":"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af","Type":"ContainerDied","Data":"9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889"} Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.104657 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" event={"ID":"8b34b3ef-6f8f-4785-afd9-aa118a3ef9af","Type":"ContainerDied","Data":"8f9494c5ffa82a680ce68bb20b0a675b3dfb162273cee72a2df964cdae9010ae"} Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.104674 4862 scope.go:117] "RemoveContainer" containerID="9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.104716 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.105509 4862 status_manager.go:851] "Failed to get status for pod" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-g5fhw\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.105691 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.105857 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.130634 4862 status_manager.go:851] "Failed to get status for pod" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-g5fhw\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.131484 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.132041 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.133250 4862 scope.go:117] "RemoveContainer" containerID="9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889" Nov 28 11:25:56 crc kubenswrapper[4862]: E1128 11:25:56.134474 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889\": container with ID starting with 9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889 not found: ID does not exist" containerID="9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.134548 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889"} err="failed to get container status \"9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889\": rpc error: code = NotFound desc = could not find container \"9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889\": container with ID starting with 9f135daa68ff771edec78ae2d5cf043f58ebe6fa7b6c56e2a55a64092afa6889 not found: ID does not exist" Nov 28 11:25:56 crc kubenswrapper[4862]: E1128 11:25:56.341967 4862 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: E1128 11:25:56.343196 4862 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: E1128 11:25:56.344025 4862 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: E1128 11:25:56.344760 4862 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: E1128 11:25:56.345397 4862 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:56 crc kubenswrapper[4862]: I1128 11:25:56.345526 4862 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 28 11:25:56 crc kubenswrapper[4862]: E1128 11:25:56.346196 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Nov 28 11:25:56 crc kubenswrapper[4862]: E1128 11:25:56.547696 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Nov 28 11:25:56 crc kubenswrapper[4862]: E1128 11:25:56.949159 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Nov 28 11:25:57 crc kubenswrapper[4862]: E1128 11:25:57.751030 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Nov 28 11:25:59 crc kubenswrapper[4862]: E1128 11:25:59.351975 4862 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="3.2s" Nov 28 11:25:59 crc kubenswrapper[4862]: I1128 11:25:59.838447 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:59 crc kubenswrapper[4862]: I1128 11:25:59.839754 4862 status_manager.go:851] "Failed to get status for pod" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-g5fhw\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:59 crc kubenswrapper[4862]: I1128 11:25:59.840082 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:59 crc kubenswrapper[4862]: I1128 11:25:59.840644 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:25:59 crc kubenswrapper[4862]: I1128 11:25:59.856281 4862 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:25:59 crc kubenswrapper[4862]: I1128 11:25:59.856325 4862 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:25:59 crc kubenswrapper[4862]: E1128 11:25:59.856931 4862 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:59 crc kubenswrapper[4862]: I1128 11:25:59.857690 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:25:59 crc kubenswrapper[4862]: W1128 11:25:59.885630 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-5297c29fbd3e091ab03a500e5d35e90ac674ebbd4d63d25931b844190ca98897 WatchSource:0}: Error finding container 5297c29fbd3e091ab03a500e5d35e90ac674ebbd4d63d25931b844190ca98897: Status 404 returned error can't find the container with id 5297c29fbd3e091ab03a500e5d35e90ac674ebbd4d63d25931b844190ca98897 Nov 28 11:26:00 crc kubenswrapper[4862]: I1128 11:26:00.135184 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5297c29fbd3e091ab03a500e5d35e90ac674ebbd4d63d25931b844190ca98897"} Nov 28 11:26:01 crc kubenswrapper[4862]: I1128 11:26:01.151798 4862 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="c985430985ea889dbf5cd6e2da21e42b5a5ba98e6a3f8f6ac542c08e94e2d63b" exitCode=0 Nov 28 11:26:01 crc kubenswrapper[4862]: I1128 11:26:01.151877 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"c985430985ea889dbf5cd6e2da21e42b5a5ba98e6a3f8f6ac542c08e94e2d63b"} Nov 28 11:26:01 crc kubenswrapper[4862]: I1128 11:26:01.152262 4862 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:26:01 crc kubenswrapper[4862]: I1128 11:26:01.152301 4862 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:26:01 crc kubenswrapper[4862]: E1128 11:26:01.152780 4862 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:26:01 crc kubenswrapper[4862]: I1128 11:26:01.152894 4862 status_manager.go:851] "Failed to get status for pod" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" pod="openshift-authentication/oauth-openshift-558db77b4-g5fhw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-g5fhw\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:26:01 crc kubenswrapper[4862]: I1128 11:26:01.154797 4862 status_manager.go:851] "Failed to get status for pod" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:26:01 crc kubenswrapper[4862]: I1128 11:26:01.155515 4862 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Nov 28 11:26:02 crc kubenswrapper[4862]: I1128 11:26:02.165616 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"363109c186dbbc625659998426812eaf09c65f7524c4b50262559dea259e46f4"} Nov 28 11:26:02 crc kubenswrapper[4862]: I1128 11:26:02.165949 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e42aea175e79b60023460ab69381d79b7dd18684ce70b15be66b887c318e9e94"} Nov 28 11:26:02 crc kubenswrapper[4862]: I1128 11:26:02.165960 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5addfcbaa9abe2b800230144a7bf2670919d537f352f505a5838cca56ea9f36a"} Nov 28 11:26:03 crc kubenswrapper[4862]: I1128 11:26:03.181130 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"52e921611d15f032958c2b8688087abac39ae357d3daf8f45963f692860b543f"} Nov 28 11:26:03 crc kubenswrapper[4862]: I1128 11:26:03.181412 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"00fe056f6a810e6a8b9678fa98626104705f656497f3ec7badde66981c49d43e"} Nov 28 11:26:03 crc kubenswrapper[4862]: I1128 11:26:03.181431 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:26:03 crc kubenswrapper[4862]: I1128 11:26:03.181466 4862 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:26:03 crc kubenswrapper[4862]: I1128 11:26:03.181493 4862 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:26:04 crc kubenswrapper[4862]: I1128 11:26:04.190425 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 11:26:04 crc kubenswrapper[4862]: I1128 11:26:04.190484 4862 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649" exitCode=1 Nov 28 11:26:04 crc kubenswrapper[4862]: I1128 11:26:04.190516 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649"} Nov 28 11:26:04 crc kubenswrapper[4862]: I1128 11:26:04.190969 4862 scope.go:117] "RemoveContainer" containerID="cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649" Nov 28 11:26:04 crc kubenswrapper[4862]: I1128 11:26:04.858071 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:26:04 crc kubenswrapper[4862]: I1128 11:26:04.858739 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:26:04 crc kubenswrapper[4862]: I1128 11:26:04.866944 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:26:05 crc kubenswrapper[4862]: I1128 11:26:05.162765 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:26:05 crc kubenswrapper[4862]: I1128 11:26:05.202317 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 11:26:05 crc kubenswrapper[4862]: I1128 11:26:05.202433 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d123564cdddc2b8ab60d80bbebe6c5bd10367717d62ff4d0ffa5e0339d1c2f22"} Nov 28 11:26:08 crc kubenswrapper[4862]: I1128 11:26:08.191593 4862 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:26:08 crc kubenswrapper[4862]: I1128 11:26:08.222595 4862 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:26:08 crc kubenswrapper[4862]: I1128 11:26:08.222637 4862 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:26:08 crc kubenswrapper[4862]: I1128 11:26:08.227827 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:26:08 crc kubenswrapper[4862]: I1128 11:26:08.230870 4862 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d9773c54-39df-417a-a5b4-92b3289e7af6" Nov 28 11:26:08 crc kubenswrapper[4862]: I1128 11:26:08.333534 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:26:09 crc kubenswrapper[4862]: I1128 11:26:09.229703 4862 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:26:09 crc kubenswrapper[4862]: I1128 11:26:09.229754 4862 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0660f02d-24c4-419c-a304-0401773eed51" Nov 28 11:26:12 crc kubenswrapper[4862]: I1128 11:26:12.862025 4862 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d9773c54-39df-417a-a5b4-92b3289e7af6" Nov 28 11:26:15 crc kubenswrapper[4862]: I1128 11:26:15.162906 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:26:15 crc kubenswrapper[4862]: I1128 11:26:15.163403 4862 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 28 11:26:15 crc kubenswrapper[4862]: I1128 11:26:15.164312 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 28 11:26:15 crc kubenswrapper[4862]: I1128 11:26:15.799484 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 11:26:18 crc kubenswrapper[4862]: I1128 11:26:18.342397 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 11:26:18 crc kubenswrapper[4862]: I1128 11:26:18.551382 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 11:26:18 crc kubenswrapper[4862]: I1128 11:26:18.606807 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 11:26:18 crc kubenswrapper[4862]: I1128 11:26:18.999647 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 11:26:20 crc kubenswrapper[4862]: I1128 11:26:20.059052 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 11:26:20 crc kubenswrapper[4862]: I1128 11:26:20.600017 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 11:26:20 crc kubenswrapper[4862]: I1128 11:26:20.698818 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 11:26:20 crc kubenswrapper[4862]: I1128 11:26:20.907753 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 11:26:21 crc kubenswrapper[4862]: I1128 11:26:21.081438 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 11:26:21 crc kubenswrapper[4862]: I1128 11:26:21.318233 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 11:26:21 crc kubenswrapper[4862]: I1128 11:26:21.374206 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 11:26:21 crc kubenswrapper[4862]: I1128 11:26:21.410192 4862 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 11:26:21 crc kubenswrapper[4862]: I1128 11:26:21.659193 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 11:26:21 crc kubenswrapper[4862]: I1128 11:26:21.707369 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 11:26:21 crc kubenswrapper[4862]: I1128 11:26:21.802565 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 11:26:21 crc kubenswrapper[4862]: I1128 11:26:21.934150 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 11:26:21 crc kubenswrapper[4862]: I1128 11:26:21.965409 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 11:26:22 crc kubenswrapper[4862]: I1128 11:26:22.044489 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 11:26:22 crc kubenswrapper[4862]: I1128 11:26:22.053986 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 11:26:22 crc kubenswrapper[4862]: I1128 11:26:22.139973 4862 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 11:26:22 crc kubenswrapper[4862]: I1128 11:26:22.334231 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 11:26:22 crc kubenswrapper[4862]: I1128 11:26:22.705263 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 11:26:22 crc kubenswrapper[4862]: I1128 11:26:22.747582 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.214710 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.284835 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.349714 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.447391 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.621122 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.631690 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.662297 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.773115 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.886784 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 11:26:23 crc kubenswrapper[4862]: I1128 11:26:23.953248 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.010848 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.075446 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.183757 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.191678 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.260925 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.272533 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.282315 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.399701 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.501120 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.643060 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.735252 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.809400 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.946145 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.948749 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 11:26:24 crc kubenswrapper[4862]: I1128 11:26:24.954013 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.029596 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.162749 4862 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.162820 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.175716 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.244294 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.305329 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.403200 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.434469 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.617504 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.634879 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.637071 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.762540 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.864943 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.878084 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.897953 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 11:26:25 crc kubenswrapper[4862]: I1128 11:26:25.910845 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.049161 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.062724 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.123851 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.125852 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.129833 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.131649 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.161490 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.174589 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.251639 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.255417 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.328119 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.333305 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.333711 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.386968 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.436741 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.438933 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.504324 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.506848 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.545543 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.571525 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.588681 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.618149 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.642939 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.646251 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.701940 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.762323 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.763853 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.852161 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.856580 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.864151 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.981239 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 11:26:26 crc kubenswrapper[4862]: I1128 11:26:26.981661 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.107966 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.134870 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.191937 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.235450 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.262110 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.296588 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.308929 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.326687 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.330268 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.337905 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.470954 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.500352 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.524223 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.562456 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.616068 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.662876 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.717202 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.743254 4862 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.845234 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.917911 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.925535 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 11:26:27 crc kubenswrapper[4862]: I1128 11:26:27.926406 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.009053 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.020876 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.051946 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.053557 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.147460 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.192776 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.199441 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.354918 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.393511 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.425398 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.502817 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.589997 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.671749 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.750931 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.789399 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.792341 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.876978 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 11:26:28 crc kubenswrapper[4862]: I1128 11:26:28.968390 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.053407 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.133316 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.187962 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.189528 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.270569 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.272120 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.349572 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.393784 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.409013 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.510675 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.575734 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.600618 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.651846 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.659504 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.684950 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.819013 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.844806 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.888172 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.920527 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.945869 4862 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.970571 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 11:26:29 crc kubenswrapper[4862]: I1128 11:26:29.992750 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.012005 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.079034 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.191869 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.200233 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.203985 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.230562 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.294121 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.329996 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.336160 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.394027 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.410253 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.479457 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.565562 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.570925 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.583653 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.613117 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.700374 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.829637 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 11:26:30 crc kubenswrapper[4862]: I1128 11:26:30.955212 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.005742 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.070842 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.204212 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.235428 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.279507 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.345838 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.378450 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.445230 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.457980 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.469568 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.636298 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.724813 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.795356 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 11:26:31 crc kubenswrapper[4862]: I1128 11:26:31.904181 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.012358 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.165226 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.218558 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.288734 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.295104 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.616502 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.674383 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.720415 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.890968 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.962391 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 11:26:32 crc kubenswrapper[4862]: I1128 11:26:32.990666 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.070185 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.140078 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.260035 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.266968 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.298642 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.398134 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.488796 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.497483 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.516850 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.529923 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.573588 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.574271 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.603455 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 11:26:33 crc kubenswrapper[4862]: I1128 11:26:33.660274 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 11:26:34 crc kubenswrapper[4862]: I1128 11:26:34.138527 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 11:26:34 crc kubenswrapper[4862]: I1128 11:26:34.183585 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 11:26:34 crc kubenswrapper[4862]: I1128 11:26:34.347796 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 11:26:34 crc kubenswrapper[4862]: I1128 11:26:34.517230 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 11:26:34 crc kubenswrapper[4862]: I1128 11:26:34.534965 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 11:26:34 crc kubenswrapper[4862]: I1128 11:26:34.592752 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 11:26:34 crc kubenswrapper[4862]: I1128 11:26:34.640551 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 11:26:34 crc kubenswrapper[4862]: I1128 11:26:34.715421 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 11:26:34 crc kubenswrapper[4862]: I1128 11:26:34.788973 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.000063 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.017957 4862 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.021015 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=47.020955797 podStartE2EDuration="47.020955797s" podCreationTimestamp="2025-11-28 11:25:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:26:08.131071038 +0000 UTC m=+245.663584949" watchObservedRunningTime="2025-11-28 11:26:35.020955797 +0000 UTC m=+272.553469748" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.026474 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g5fhw","openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.026560 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.026590 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cph87"] Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.026921 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cph87" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerName="registry-server" containerID="cri-o://15cdd6a93a5ae1894530992e43a38d30e304f4771e10c25cb03ae9617b523d5f" gracePeriod=2 Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.034070 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.071558 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=27.0715279 podStartE2EDuration="27.0715279s" podCreationTimestamp="2025-11-28 11:26:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:26:35.065782167 +0000 UTC m=+272.598296118" watchObservedRunningTime="2025-11-28 11:26:35.0715279 +0000 UTC m=+272.604041861" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.114607 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.152962 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.162528 4862 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.162605 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.162673 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.163571 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"d123564cdddc2b8ab60d80bbebe6c5bd10367717d62ff4d0ffa5e0339d1c2f22"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.163766 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://d123564cdddc2b8ab60d80bbebe6c5bd10367717d62ff4d0ffa5e0339d1c2f22" gracePeriod=30 Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.180152 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.184865 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.600860 4862 generic.go:334] "Generic (PLEG): container finished" podID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerID="15cdd6a93a5ae1894530992e43a38d30e304f4771e10c25cb03ae9617b523d5f" exitCode=0 Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.600909 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cph87" event={"ID":"22c7db45-0f31-4bc7-b93d-4470c51434b1","Type":"ContainerDied","Data":"15cdd6a93a5ae1894530992e43a38d30e304f4771e10c25cb03ae9617b523d5f"} Nov 28 11:26:35 crc kubenswrapper[4862]: I1128 11:26:35.632521 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.087741 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.102559 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.175828 4862 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.265023 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.418027 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-catalog-content\") pod \"22c7db45-0f31-4bc7-b93d-4470c51434b1\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.418154 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djjls\" (UniqueName: \"kubernetes.io/projected/22c7db45-0f31-4bc7-b93d-4470c51434b1-kube-api-access-djjls\") pod \"22c7db45-0f31-4bc7-b93d-4470c51434b1\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.418217 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-utilities\") pod \"22c7db45-0f31-4bc7-b93d-4470c51434b1\" (UID: \"22c7db45-0f31-4bc7-b93d-4470c51434b1\") " Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.420410 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-utilities" (OuterVolumeSpecName: "utilities") pod "22c7db45-0f31-4bc7-b93d-4470c51434b1" (UID: "22c7db45-0f31-4bc7-b93d-4470c51434b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.429380 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c7db45-0f31-4bc7-b93d-4470c51434b1-kube-api-access-djjls" (OuterVolumeSpecName: "kube-api-access-djjls") pod "22c7db45-0f31-4bc7-b93d-4470c51434b1" (UID: "22c7db45-0f31-4bc7-b93d-4470c51434b1"). InnerVolumeSpecName "kube-api-access-djjls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.507400 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22c7db45-0f31-4bc7-b93d-4470c51434b1" (UID: "22c7db45-0f31-4bc7-b93d-4470c51434b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.520933 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.521197 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djjls\" (UniqueName: \"kubernetes.io/projected/22c7db45-0f31-4bc7-b93d-4470c51434b1-kube-api-access-djjls\") on node \"crc\" DevicePath \"\"" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.521367 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c7db45-0f31-4bc7-b93d-4470c51434b1-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.536977 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.623845 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cph87" event={"ID":"22c7db45-0f31-4bc7-b93d-4470c51434b1","Type":"ContainerDied","Data":"0a91f6cde582434aa43401b24dc91ae2a291732b32400964ab4fc090cb6075fb"} Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.623957 4862 scope.go:117] "RemoveContainer" containerID="15cdd6a93a5ae1894530992e43a38d30e304f4771e10c25cb03ae9617b523d5f" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.624011 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cph87" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.665982 4862 scope.go:117] "RemoveContainer" containerID="31b511f511059bf076368b01a7cf350621fef7692bcc48216e0afd850541b00f" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.699404 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cph87"] Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.704474 4862 scope.go:117] "RemoveContainer" containerID="bda8bae1495e604ef029a0ecb45cbe5ce2d59fe254975608d8370b3d5ba1c77a" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.705915 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cph87"] Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.855449 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" path="/var/lib/kubelet/pods/22c7db45-0f31-4bc7-b93d-4470c51434b1/volumes" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.856804 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" path="/var/lib/kubelet/pods/8b34b3ef-6f8f-4785-afd9-aa118a3ef9af/volumes" Nov 28 11:26:36 crc kubenswrapper[4862]: I1128 11:26:36.957159 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 11:26:38 crc kubenswrapper[4862]: I1128 11:26:38.956700 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.703862 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-64b5cd4787-z2zb2"] Nov 28 11:26:41 crc kubenswrapper[4862]: E1128 11:26:41.706706 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" containerName="oauth-openshift" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.706756 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" containerName="oauth-openshift" Nov 28 11:26:41 crc kubenswrapper[4862]: E1128 11:26:41.706782 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerName="extract-content" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.706796 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerName="extract-content" Nov 28 11:26:41 crc kubenswrapper[4862]: E1128 11:26:41.706823 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerName="extract-utilities" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.706841 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerName="extract-utilities" Nov 28 11:26:41 crc kubenswrapper[4862]: E1128 11:26:41.706870 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" containerName="installer" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.706887 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" containerName="installer" Nov 28 11:26:41 crc kubenswrapper[4862]: E1128 11:26:41.706913 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerName="registry-server" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.706929 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerName="registry-server" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.707198 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b34b3ef-6f8f-4785-afd9-aa118a3ef9af" containerName="oauth-openshift" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.707227 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b644bbd7-4723-4b57-9844-d956bbbb964c" containerName="installer" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.707248 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="22c7db45-0f31-4bc7-b93d-4470c51434b1" containerName="registry-server" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.709168 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.715716 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.717040 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.717171 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.717359 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.717402 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.717527 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.718052 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.718239 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.718963 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.718964 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.719370 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.724840 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.738353 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64b5cd4787-z2zb2"] Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.739195 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.741449 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.750932 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.799652 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.799730 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.799783 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.799870 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.799918 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t9x4\" (UniqueName: \"kubernetes.io/projected/41f0d376-8574-4109-a9a6-c82ad871b2b7-kube-api-access-4t9x4\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.800002 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.800053 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-audit-policies\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.800159 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-session\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.800185 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.800417 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-template-error\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.800518 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/41f0d376-8574-4109-a9a6-c82ad871b2b7-audit-dir\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.800558 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-router-certs\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.800599 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-template-login\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.800725 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-service-ca\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.901975 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-router-certs\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902143 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-template-login\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902240 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-service-ca\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902325 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902382 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902440 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902509 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902577 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t9x4\" (UniqueName: \"kubernetes.io/projected/41f0d376-8574-4109-a9a6-c82ad871b2b7-kube-api-access-4t9x4\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902660 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902707 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-audit-policies\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902772 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-session\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902810 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902887 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-template-error\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.902948 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/41f0d376-8574-4109-a9a6-c82ad871b2b7-audit-dir\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.903086 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/41f0d376-8574-4109-a9a6-c82ad871b2b7-audit-dir\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.904455 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-service-ca\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.904624 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.905500 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.905796 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/41f0d376-8574-4109-a9a6-c82ad871b2b7-audit-policies\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.911649 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-session\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.911681 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-template-login\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.911713 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.912029 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.912750 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-template-error\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.915488 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-router-certs\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.916272 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.917227 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/41f0d376-8574-4109-a9a6-c82ad871b2b7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:41 crc kubenswrapper[4862]: I1128 11:26:41.938916 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t9x4\" (UniqueName: \"kubernetes.io/projected/41f0d376-8574-4109-a9a6-c82ad871b2b7-kube-api-access-4t9x4\") pod \"oauth-openshift-64b5cd4787-z2zb2\" (UID: \"41f0d376-8574-4109-a9a6-c82ad871b2b7\") " pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:42 crc kubenswrapper[4862]: I1128 11:26:42.051844 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:42 crc kubenswrapper[4862]: I1128 11:26:42.089490 4862 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 11:26:42 crc kubenswrapper[4862]: I1128 11:26:42.090320 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565" gracePeriod=5 Nov 28 11:26:42 crc kubenswrapper[4862]: I1128 11:26:42.290434 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64b5cd4787-z2zb2"] Nov 28 11:26:42 crc kubenswrapper[4862]: I1128 11:26:42.672623 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" event={"ID":"41f0d376-8574-4109-a9a6-c82ad871b2b7","Type":"ContainerStarted","Data":"52841eb344b57079918370d90d4ca9a62756c59b11f5f3dc33f3611e1e2b7368"} Nov 28 11:26:42 crc kubenswrapper[4862]: I1128 11:26:42.672834 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" event={"ID":"41f0d376-8574-4109-a9a6-c82ad871b2b7","Type":"ContainerStarted","Data":"f0902fea132cd6f2af85a709df2185c708c7b38c629a6793a7f60b0811c21add"} Nov 28 11:26:42 crc kubenswrapper[4862]: I1128 11:26:42.672884 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:42 crc kubenswrapper[4862]: I1128 11:26:42.726370 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" podStartSLOduration=72.72634755 podStartE2EDuration="1m12.72634755s" podCreationTimestamp="2025-11-28 11:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:26:42.719769186 +0000 UTC m=+280.252283147" watchObservedRunningTime="2025-11-28 11:26:42.72634755 +0000 UTC m=+280.258861481" Nov 28 11:26:43 crc kubenswrapper[4862]: I1128 11:26:43.092379 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-64b5cd4787-z2zb2" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.703631 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.704115 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.712185 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.712277 4862 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565" exitCode=137 Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.712363 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.712401 4862 scope.go:117] "RemoveContainer" containerID="b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.737695 4862 scope.go:117] "RemoveContainer" containerID="b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565" Nov 28 11:26:47 crc kubenswrapper[4862]: E1128 11:26:47.738278 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565\": container with ID starting with b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565 not found: ID does not exist" containerID="b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.738316 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565"} err="failed to get container status \"b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565\": rpc error: code = NotFound desc = could not find container \"b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565\": container with ID starting with b3a9b79f983494d2124f771c9a541fd655f315975413d4fce37eca681f1d7565 not found: ID does not exist" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.801549 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.801687 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.801767 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.801877 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.801917 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.801875 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.802037 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.802211 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.802305 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.802617 4862 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.802648 4862 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.802664 4862 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.802682 4862 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.816452 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:26:47 crc kubenswrapper[4862]: I1128 11:26:47.904685 4862 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 11:26:48 crc kubenswrapper[4862]: I1128 11:26:48.850711 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 28 11:26:48 crc kubenswrapper[4862]: I1128 11:26:48.851321 4862 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 28 11:26:48 crc kubenswrapper[4862]: I1128 11:26:48.866514 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 11:26:48 crc kubenswrapper[4862]: I1128 11:26:48.866568 4862 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="12b3ccdf-040b-44f2-884a-57ecc7f86aa0" Nov 28 11:26:48 crc kubenswrapper[4862]: I1128 11:26:48.874178 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 11:26:48 crc kubenswrapper[4862]: I1128 11:26:48.874427 4862 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="12b3ccdf-040b-44f2-884a-57ecc7f86aa0" Nov 28 11:26:55 crc kubenswrapper[4862]: I1128 11:26:55.771577 4862 generic.go:334] "Generic (PLEG): container finished" podID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerID="7ce732110dabdb015cb436adb7d4492085119d3ed7379ec98d25dff4366d4872" exitCode=0 Nov 28 11:26:55 crc kubenswrapper[4862]: I1128 11:26:55.771703 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" event={"ID":"3b577ecb-1896-4699-bd7e-33929f2cf9a3","Type":"ContainerDied","Data":"7ce732110dabdb015cb436adb7d4492085119d3ed7379ec98d25dff4366d4872"} Nov 28 11:26:55 crc kubenswrapper[4862]: I1128 11:26:55.773341 4862 scope.go:117] "RemoveContainer" containerID="7ce732110dabdb015cb436adb7d4492085119d3ed7379ec98d25dff4366d4872" Nov 28 11:26:56 crc kubenswrapper[4862]: I1128 11:26:56.781243 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" event={"ID":"3b577ecb-1896-4699-bd7e-33929f2cf9a3","Type":"ContainerStarted","Data":"2d7ab602e21fb4d1d4b1581a5674379862206b9bfe261952d3f640e5e4281f59"} Nov 28 11:26:56 crc kubenswrapper[4862]: I1128 11:26:56.782127 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:26:56 crc kubenswrapper[4862]: I1128 11:26:56.785451 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:27:05 crc kubenswrapper[4862]: I1128 11:27:05.844962 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Nov 28 11:27:05 crc kubenswrapper[4862]: I1128 11:27:05.852590 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 11:27:05 crc kubenswrapper[4862]: I1128 11:27:05.852683 4862 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="d123564cdddc2b8ab60d80bbebe6c5bd10367717d62ff4d0ffa5e0339d1c2f22" exitCode=137 Nov 28 11:27:05 crc kubenswrapper[4862]: I1128 11:27:05.852738 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"d123564cdddc2b8ab60d80bbebe6c5bd10367717d62ff4d0ffa5e0339d1c2f22"} Nov 28 11:27:05 crc kubenswrapper[4862]: I1128 11:27:05.852817 4862 scope.go:117] "RemoveContainer" containerID="cdfe065d7855667f1c254be88b6101731f853d0ccb18fb19f6468fd80a415649" Nov 28 11:27:06 crc kubenswrapper[4862]: I1128 11:27:06.862432 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Nov 28 11:27:06 crc kubenswrapper[4862]: I1128 11:27:06.866787 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c09d4e5a05a9c84c328b1c20dd943a2ddcb03724bafee1839533bcf2aa90eb91"} Nov 28 11:27:08 crc kubenswrapper[4862]: I1128 11:27:08.333360 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:27:15 crc kubenswrapper[4862]: I1128 11:27:15.162534 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:27:15 crc kubenswrapper[4862]: I1128 11:27:15.166715 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:27:15 crc kubenswrapper[4862]: I1128 11:27:15.928616 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.291198 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhl4q"] Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.292311 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" podUID="9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" containerName="controller-manager" containerID="cri-o://3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91" gracePeriod=30 Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.295374 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls"] Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.295637 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" podUID="ee1f3625-a084-4e0d-99af-627305b0f8f2" containerName="route-controller-manager" containerID="cri-o://13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c" gracePeriod=30 Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.731870 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.791780 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.871737 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee1f3625-a084-4e0d-99af-627305b0f8f2-serving-cert\") pod \"ee1f3625-a084-4e0d-99af-627305b0f8f2\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.871829 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-config\") pod \"ee1f3625-a084-4e0d-99af-627305b0f8f2\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.871877 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-client-ca\") pod \"ee1f3625-a084-4e0d-99af-627305b0f8f2\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.872010 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sscpn\" (UniqueName: \"kubernetes.io/projected/ee1f3625-a084-4e0d-99af-627305b0f8f2-kube-api-access-sscpn\") pod \"ee1f3625-a084-4e0d-99af-627305b0f8f2\" (UID: \"ee1f3625-a084-4e0d-99af-627305b0f8f2\") " Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.873795 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-client-ca" (OuterVolumeSpecName: "client-ca") pod "ee1f3625-a084-4e0d-99af-627305b0f8f2" (UID: "ee1f3625-a084-4e0d-99af-627305b0f8f2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.873935 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-config" (OuterVolumeSpecName: "config") pod "ee1f3625-a084-4e0d-99af-627305b0f8f2" (UID: "ee1f3625-a084-4e0d-99af-627305b0f8f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.880724 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee1f3625-a084-4e0d-99af-627305b0f8f2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ee1f3625-a084-4e0d-99af-627305b0f8f2" (UID: "ee1f3625-a084-4e0d-99af-627305b0f8f2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.881539 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee1f3625-a084-4e0d-99af-627305b0f8f2-kube-api-access-sscpn" (OuterVolumeSpecName: "kube-api-access-sscpn") pod "ee1f3625-a084-4e0d-99af-627305b0f8f2" (UID: "ee1f3625-a084-4e0d-99af-627305b0f8f2"). InnerVolumeSpecName "kube-api-access-sscpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.971668 4862 generic.go:334] "Generic (PLEG): container finished" podID="ee1f3625-a084-4e0d-99af-627305b0f8f2" containerID="13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c" exitCode=0 Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.971746 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" event={"ID":"ee1f3625-a084-4e0d-99af-627305b0f8f2","Type":"ContainerDied","Data":"13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c"} Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.971835 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" event={"ID":"ee1f3625-a084-4e0d-99af-627305b0f8f2","Type":"ContainerDied","Data":"097b4b645db34dd3616fe180841b937e52c189be76d11fb9eada62279fdb38e9"} Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.971841 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.971865 4862 scope.go:117] "RemoveContainer" containerID="13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973419 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-serving-cert\") pod \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973517 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-proxy-ca-bundles\") pod \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973559 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-config\") pod \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973624 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-client-ca\") pod \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973648 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" event={"ID":"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560","Type":"ContainerDied","Data":"3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91"} Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973687 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqtx8\" (UniqueName: \"kubernetes.io/projected/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-kube-api-access-qqtx8\") pod \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\" (UID: \"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560\") " Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973796 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973609 4862 generic.go:334] "Generic (PLEG): container finished" podID="9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" containerID="3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91" exitCode=0 Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973899 4862 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973915 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sscpn\" (UniqueName: \"kubernetes.io/projected/ee1f3625-a084-4e0d-99af-627305b0f8f2-kube-api-access-sscpn\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973930 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee1f3625-a084-4e0d-99af-627305b0f8f2-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.973926 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bhl4q" event={"ID":"9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560","Type":"ContainerDied","Data":"834909e4c3997062db18068324837fc15707986f43331043cfc6cb939bdfbc66"} Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.974313 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1f3625-a084-4e0d-99af-627305b0f8f2-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.974319 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" (UID: "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.974501 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-client-ca" (OuterVolumeSpecName: "client-ca") pod "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" (UID: "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.974716 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-config" (OuterVolumeSpecName: "config") pod "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" (UID: "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.977264 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" (UID: "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.981564 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-kube-api-access-qqtx8" (OuterVolumeSpecName: "kube-api-access-qqtx8") pod "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" (UID: "9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560"). InnerVolumeSpecName "kube-api-access-qqtx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.994911 4862 scope.go:117] "RemoveContainer" containerID="13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c" Nov 28 11:27:22 crc kubenswrapper[4862]: E1128 11:27:22.995461 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c\": container with ID starting with 13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c not found: ID does not exist" containerID="13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.995514 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c"} err="failed to get container status \"13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c\": rpc error: code = NotFound desc = could not find container \"13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c\": container with ID starting with 13dfb5c61616a23a8b0940af3600f7efd8e5c0bb64be6d82d17f00b16c6edc8c not found: ID does not exist" Nov 28 11:27:22 crc kubenswrapper[4862]: I1128 11:27:22.995553 4862 scope.go:117] "RemoveContainer" containerID="3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.013975 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls"] Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.020295 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5qls"] Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.024305 4862 scope.go:117] "RemoveContainer" containerID="3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91" Nov 28 11:27:23 crc kubenswrapper[4862]: E1128 11:27:23.025003 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91\": container with ID starting with 3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91 not found: ID does not exist" containerID="3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.025045 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91"} err="failed to get container status \"3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91\": rpc error: code = NotFound desc = could not find container \"3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91\": container with ID starting with 3a5391157ea61a163b2a20d517c29203104efd2e46414d73ed2d9a43b9139a91 not found: ID does not exist" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.075350 4862 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.075372 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.075381 4862 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.075391 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqtx8\" (UniqueName: \"kubernetes.io/projected/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-kube-api-access-qqtx8\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.075401 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.303390 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhl4q"] Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.306782 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhl4q"] Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.720870 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc"] Nov 28 11:27:23 crc kubenswrapper[4862]: E1128 11:27:23.721309 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.721337 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 11:27:23 crc kubenswrapper[4862]: E1128 11:27:23.721389 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee1f3625-a084-4e0d-99af-627305b0f8f2" containerName="route-controller-manager" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.721407 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee1f3625-a084-4e0d-99af-627305b0f8f2" containerName="route-controller-manager" Nov 28 11:27:23 crc kubenswrapper[4862]: E1128 11:27:23.721431 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" containerName="controller-manager" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.721448 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" containerName="controller-manager" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.721642 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" containerName="controller-manager" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.721688 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.721713 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee1f3625-a084-4e0d-99af-627305b0f8f2" containerName="route-controller-manager" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.722479 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.725276 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.726259 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.726479 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.727410 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc"] Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.727809 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.728575 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.731687 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.732957 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.733766 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.733875 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.734738 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.734761 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.734764 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.734832 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.746021 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc"] Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.788690 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-config\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.788796 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-client-ca\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.788870 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-config\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.788930 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hxz9\" (UniqueName: \"kubernetes.io/projected/f913d7e8-7257-4370-b558-2324f00d051f-kube-api-access-5hxz9\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.788994 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cdd6814-97db-4391-8d2e-e7473625da64-serving-cert\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.789070 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5nr5\" (UniqueName: \"kubernetes.io/projected/5cdd6814-97db-4391-8d2e-e7473625da64-kube-api-access-h5nr5\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.789183 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-proxy-ca-bundles\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.789242 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-client-ca\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.789308 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f913d7e8-7257-4370-b558-2324f00d051f-serving-cert\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.791747 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.796176 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc"] Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.890802 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-proxy-ca-bundles\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.890872 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-client-ca\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.890909 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f913d7e8-7257-4370-b558-2324f00d051f-serving-cert\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.891031 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-config\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.891070 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-client-ca\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.891153 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-config\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.891187 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hxz9\" (UniqueName: \"kubernetes.io/projected/f913d7e8-7257-4370-b558-2324f00d051f-kube-api-access-5hxz9\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.891227 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cdd6814-97db-4391-8d2e-e7473625da64-serving-cert\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.891306 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5nr5\" (UniqueName: \"kubernetes.io/projected/5cdd6814-97db-4391-8d2e-e7473625da64-kube-api-access-h5nr5\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.892271 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-client-ca\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.892420 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-config\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.893308 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-client-ca\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.893379 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-proxy-ca-bundles\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.893500 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-config\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.898877 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cdd6814-97db-4391-8d2e-e7473625da64-serving-cert\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.899269 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f913d7e8-7257-4370-b558-2324f00d051f-serving-cert\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.917778 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hxz9\" (UniqueName: \"kubernetes.io/projected/f913d7e8-7257-4370-b558-2324f00d051f-kube-api-access-5hxz9\") pod \"route-controller-manager-74d76d5567-8m8jc\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:23 crc kubenswrapper[4862]: I1128 11:27:23.919857 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5nr5\" (UniqueName: \"kubernetes.io/projected/5cdd6814-97db-4391-8d2e-e7473625da64-kube-api-access-h5nr5\") pod \"controller-manager-5d4b59f8f9-n6mvc\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.042986 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.102001 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.308741 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc"] Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.398202 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc"] Nov 28 11:27:24 crc kubenswrapper[4862]: W1128 11:27:24.411392 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf913d7e8_7257_4370_b558_2324f00d051f.slice/crio-ccd33c63366233c235b76ceccc26e231c2cf20c3a223ed2ec3712bc35a949761 WatchSource:0}: Error finding container ccd33c63366233c235b76ceccc26e231c2cf20c3a223ed2ec3712bc35a949761: Status 404 returned error can't find the container with id ccd33c63366233c235b76ceccc26e231c2cf20c3a223ed2ec3712bc35a949761 Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.845295 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560" path="/var/lib/kubelet/pods/9b91d7d3-c7c4-4cdb-9fb7-de8c068e9560/volumes" Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.846877 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee1f3625-a084-4e0d-99af-627305b0f8f2" path="/var/lib/kubelet/pods/ee1f3625-a084-4e0d-99af-627305b0f8f2/volumes" Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.996656 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" event={"ID":"f913d7e8-7257-4370-b558-2324f00d051f","Type":"ContainerStarted","Data":"a2ac6e305f45234c3221bd262c18413a5d8d0eef53f3d02c7876502b0e2543a6"} Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.996718 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" event={"ID":"f913d7e8-7257-4370-b558-2324f00d051f","Type":"ContainerStarted","Data":"ccd33c63366233c235b76ceccc26e231c2cf20c3a223ed2ec3712bc35a949761"} Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.997022 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.998725 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" event={"ID":"5cdd6814-97db-4391-8d2e-e7473625da64","Type":"ContainerStarted","Data":"8363f1af0a6a7cc95132e4df873c28f9288bae43eda42c77241f5cfd5c251401"} Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.998776 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" event={"ID":"5cdd6814-97db-4391-8d2e-e7473625da64","Type":"ContainerStarted","Data":"a88c7cb6f2f0f13c24bccb27bff649ba8d90d514abf69a62e16822b5e5f13846"} Nov 28 11:27:24 crc kubenswrapper[4862]: I1128 11:27:24.998949 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:25 crc kubenswrapper[4862]: I1128 11:27:25.003310 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:25 crc kubenswrapper[4862]: I1128 11:27:25.019351 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:25 crc kubenswrapper[4862]: I1128 11:27:25.037613 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" podStartSLOduration=3.037585872 podStartE2EDuration="3.037585872s" podCreationTimestamp="2025-11-28 11:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:27:25.026166235 +0000 UTC m=+322.558680156" watchObservedRunningTime="2025-11-28 11:27:25.037585872 +0000 UTC m=+322.570099803" Nov 28 11:27:25 crc kubenswrapper[4862]: I1128 11:27:25.074219 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" podStartSLOduration=3.074197553 podStartE2EDuration="3.074197553s" podCreationTimestamp="2025-11-28 11:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:27:25.070256267 +0000 UTC m=+322.602770218" watchObservedRunningTime="2025-11-28 11:27:25.074197553 +0000 UTC m=+322.606711474" Nov 28 11:27:32 crc kubenswrapper[4862]: I1128 11:27:32.112780 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc"] Nov 28 11:27:32 crc kubenswrapper[4862]: I1128 11:27:32.113942 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" podUID="5cdd6814-97db-4391-8d2e-e7473625da64" containerName="controller-manager" containerID="cri-o://8363f1af0a6a7cc95132e4df873c28f9288bae43eda42c77241f5cfd5c251401" gracePeriod=30 Nov 28 11:27:32 crc kubenswrapper[4862]: I1128 11:27:32.127478 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc"] Nov 28 11:27:32 crc kubenswrapper[4862]: I1128 11:27:32.127748 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" podUID="f913d7e8-7257-4370-b558-2324f00d051f" containerName="route-controller-manager" containerID="cri-o://a2ac6e305f45234c3221bd262c18413a5d8d0eef53f3d02c7876502b0e2543a6" gracePeriod=30 Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.051066 4862 generic.go:334] "Generic (PLEG): container finished" podID="f913d7e8-7257-4370-b558-2324f00d051f" containerID="a2ac6e305f45234c3221bd262c18413a5d8d0eef53f3d02c7876502b0e2543a6" exitCode=0 Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.051407 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" event={"ID":"f913d7e8-7257-4370-b558-2324f00d051f","Type":"ContainerDied","Data":"a2ac6e305f45234c3221bd262c18413a5d8d0eef53f3d02c7876502b0e2543a6"} Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.053174 4862 generic.go:334] "Generic (PLEG): container finished" podID="5cdd6814-97db-4391-8d2e-e7473625da64" containerID="8363f1af0a6a7cc95132e4df873c28f9288bae43eda42c77241f5cfd5c251401" exitCode=0 Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.053198 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" event={"ID":"5cdd6814-97db-4391-8d2e-e7473625da64","Type":"ContainerDied","Data":"8363f1af0a6a7cc95132e4df873c28f9288bae43eda42c77241f5cfd5c251401"} Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.215834 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.244135 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc"] Nov 28 11:27:33 crc kubenswrapper[4862]: E1128 11:27:33.244657 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f913d7e8-7257-4370-b558-2324f00d051f" containerName="route-controller-manager" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.244669 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f913d7e8-7257-4370-b558-2324f00d051f" containerName="route-controller-manager" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.244783 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f913d7e8-7257-4370-b558-2324f00d051f" containerName="route-controller-manager" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.245155 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.263935 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc"] Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.313546 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.361152 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-client-ca\") pod \"f913d7e8-7257-4370-b558-2324f00d051f\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.361232 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hxz9\" (UniqueName: \"kubernetes.io/projected/f913d7e8-7257-4370-b558-2324f00d051f-kube-api-access-5hxz9\") pod \"f913d7e8-7257-4370-b558-2324f00d051f\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.361298 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f913d7e8-7257-4370-b558-2324f00d051f-serving-cert\") pod \"f913d7e8-7257-4370-b558-2324f00d051f\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.361325 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-config\") pod \"f913d7e8-7257-4370-b558-2324f00d051f\" (UID: \"f913d7e8-7257-4370-b558-2324f00d051f\") " Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.361485 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v4qt\" (UniqueName: \"kubernetes.io/projected/1873af10-25b9-4d13-83eb-c9b3fc800aeb-kube-api-access-8v4qt\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.361511 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1873af10-25b9-4d13-83eb-c9b3fc800aeb-config\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.361619 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1873af10-25b9-4d13-83eb-c9b3fc800aeb-client-ca\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.361898 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1873af10-25b9-4d13-83eb-c9b3fc800aeb-serving-cert\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.362411 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-client-ca" (OuterVolumeSpecName: "client-ca") pod "f913d7e8-7257-4370-b558-2324f00d051f" (UID: "f913d7e8-7257-4370-b558-2324f00d051f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.362530 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-config" (OuterVolumeSpecName: "config") pod "f913d7e8-7257-4370-b558-2324f00d051f" (UID: "f913d7e8-7257-4370-b558-2324f00d051f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.368984 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f913d7e8-7257-4370-b558-2324f00d051f-kube-api-access-5hxz9" (OuterVolumeSpecName: "kube-api-access-5hxz9") pod "f913d7e8-7257-4370-b558-2324f00d051f" (UID: "f913d7e8-7257-4370-b558-2324f00d051f"). InnerVolumeSpecName "kube-api-access-5hxz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.369636 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f913d7e8-7257-4370-b558-2324f00d051f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f913d7e8-7257-4370-b558-2324f00d051f" (UID: "f913d7e8-7257-4370-b558-2324f00d051f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.462594 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cdd6814-97db-4391-8d2e-e7473625da64-serving-cert\") pod \"5cdd6814-97db-4391-8d2e-e7473625da64\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.462673 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-client-ca\") pod \"5cdd6814-97db-4391-8d2e-e7473625da64\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.462730 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5nr5\" (UniqueName: \"kubernetes.io/projected/5cdd6814-97db-4391-8d2e-e7473625da64-kube-api-access-h5nr5\") pod \"5cdd6814-97db-4391-8d2e-e7473625da64\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.462775 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-config\") pod \"5cdd6814-97db-4391-8d2e-e7473625da64\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.462805 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-proxy-ca-bundles\") pod \"5cdd6814-97db-4391-8d2e-e7473625da64\" (UID: \"5cdd6814-97db-4391-8d2e-e7473625da64\") " Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.463019 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1873af10-25b9-4d13-83eb-c9b3fc800aeb-client-ca\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.463078 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1873af10-25b9-4d13-83eb-c9b3fc800aeb-serving-cert\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.463217 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v4qt\" (UniqueName: \"kubernetes.io/projected/1873af10-25b9-4d13-83eb-c9b3fc800aeb-kube-api-access-8v4qt\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.463248 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1873af10-25b9-4d13-83eb-c9b3fc800aeb-config\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.463302 4862 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.463322 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hxz9\" (UniqueName: \"kubernetes.io/projected/f913d7e8-7257-4370-b558-2324f00d051f-kube-api-access-5hxz9\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.463360 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f913d7e8-7257-4370-b558-2324f00d051f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.463377 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f913d7e8-7257-4370-b558-2324f00d051f-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.463910 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5cdd6814-97db-4391-8d2e-e7473625da64" (UID: "5cdd6814-97db-4391-8d2e-e7473625da64"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.464027 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-client-ca" (OuterVolumeSpecName: "client-ca") pod "5cdd6814-97db-4391-8d2e-e7473625da64" (UID: "5cdd6814-97db-4391-8d2e-e7473625da64"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.464182 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-config" (OuterVolumeSpecName: "config") pod "5cdd6814-97db-4391-8d2e-e7473625da64" (UID: "5cdd6814-97db-4391-8d2e-e7473625da64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.464477 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1873af10-25b9-4d13-83eb-c9b3fc800aeb-client-ca\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.465176 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1873af10-25b9-4d13-83eb-c9b3fc800aeb-config\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.467404 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1873af10-25b9-4d13-83eb-c9b3fc800aeb-serving-cert\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.470316 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cdd6814-97db-4391-8d2e-e7473625da64-kube-api-access-h5nr5" (OuterVolumeSpecName: "kube-api-access-h5nr5") pod "5cdd6814-97db-4391-8d2e-e7473625da64" (UID: "5cdd6814-97db-4391-8d2e-e7473625da64"). InnerVolumeSpecName "kube-api-access-h5nr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.470391 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cdd6814-97db-4391-8d2e-e7473625da64-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5cdd6814-97db-4391-8d2e-e7473625da64" (UID: "5cdd6814-97db-4391-8d2e-e7473625da64"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.491289 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v4qt\" (UniqueName: \"kubernetes.io/projected/1873af10-25b9-4d13-83eb-c9b3fc800aeb-kube-api-access-8v4qt\") pod \"route-controller-manager-6884b9b445-8bwmc\" (UID: \"1873af10-25b9-4d13-83eb-c9b3fc800aeb\") " pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.559217 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.564386 4862 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.564471 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5nr5\" (UniqueName: \"kubernetes.io/projected/5cdd6814-97db-4391-8d2e-e7473625da64-kube-api-access-h5nr5\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.564486 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.564498 4862 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cdd6814-97db-4391-8d2e-e7473625da64-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:33 crc kubenswrapper[4862]: I1128 11:27:33.564509 4862 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cdd6814-97db-4391-8d2e-e7473625da64-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.039404 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc"] Nov 28 11:27:34 crc kubenswrapper[4862]: W1128 11:27:34.049689 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1873af10_25b9_4d13_83eb_c9b3fc800aeb.slice/crio-17206c51861f255bab734f8b5b66b93aca76b5dbbaca4f218798b14e1f0a9f9f WatchSource:0}: Error finding container 17206c51861f255bab734f8b5b66b93aca76b5dbbaca4f218798b14e1f0a9f9f: Status 404 returned error can't find the container with id 17206c51861f255bab734f8b5b66b93aca76b5dbbaca4f218798b14e1f0a9f9f Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.068579 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" event={"ID":"5cdd6814-97db-4391-8d2e-e7473625da64","Type":"ContainerDied","Data":"a88c7cb6f2f0f13c24bccb27bff649ba8d90d514abf69a62e16822b5e5f13846"} Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.068690 4862 scope.go:117] "RemoveContainer" containerID="8363f1af0a6a7cc95132e4df873c28f9288bae43eda42c77241f5cfd5c251401" Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.068875 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc" Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.071666 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" event={"ID":"1873af10-25b9-4d13-83eb-c9b3fc800aeb","Type":"ContainerStarted","Data":"17206c51861f255bab734f8b5b66b93aca76b5dbbaca4f218798b14e1f0a9f9f"} Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.075043 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" event={"ID":"f913d7e8-7257-4370-b558-2324f00d051f","Type":"ContainerDied","Data":"ccd33c63366233c235b76ceccc26e231c2cf20c3a223ed2ec3712bc35a949761"} Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.075116 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc" Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.096973 4862 scope.go:117] "RemoveContainer" containerID="a2ac6e305f45234c3221bd262c18413a5d8d0eef53f3d02c7876502b0e2543a6" Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.113991 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc"] Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.120053 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74d76d5567-8m8jc"] Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.128271 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc"] Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.132011 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5d4b59f8f9-n6mvc"] Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.849617 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cdd6814-97db-4391-8d2e-e7473625da64" path="/var/lib/kubelet/pods/5cdd6814-97db-4391-8d2e-e7473625da64/volumes" Nov 28 11:27:34 crc kubenswrapper[4862]: I1128 11:27:34.852025 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f913d7e8-7257-4370-b558-2324f00d051f" path="/var/lib/kubelet/pods/f913d7e8-7257-4370-b558-2324f00d051f/volumes" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.080497 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" event={"ID":"1873af10-25b9-4d13-83eb-c9b3fc800aeb","Type":"ContainerStarted","Data":"98e8f95d7c885f26acf05190b7d09c422f50c924d4dd5031cd1beb2f2e52e446"} Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.080767 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.088495 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.102924 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6884b9b445-8bwmc" podStartSLOduration=3.102900754 podStartE2EDuration="3.102900754s" podCreationTimestamp="2025-11-28 11:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:27:35.097478947 +0000 UTC m=+332.629992888" watchObservedRunningTime="2025-11-28 11:27:35.102900754 +0000 UTC m=+332.635414695" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.727166 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-559bb546ff-8kkhz"] Nov 28 11:27:35 crc kubenswrapper[4862]: E1128 11:27:35.727387 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cdd6814-97db-4391-8d2e-e7473625da64" containerName="controller-manager" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.727399 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cdd6814-97db-4391-8d2e-e7473625da64" containerName="controller-manager" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.727503 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cdd6814-97db-4391-8d2e-e7473625da64" containerName="controller-manager" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.727857 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.733979 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.734219 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.734440 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.734594 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.734697 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.734798 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.749455 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.752653 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-559bb546ff-8kkhz"] Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.871159 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbfd688a-ee0f-495b-a915-f0345172d2cc-config\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.871283 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbfd688a-ee0f-495b-a915-f0345172d2cc-serving-cert\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.871338 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbfd688a-ee0f-495b-a915-f0345172d2cc-proxy-ca-bundles\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.871378 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8l4m\" (UniqueName: \"kubernetes.io/projected/cbfd688a-ee0f-495b-a915-f0345172d2cc-kube-api-access-r8l4m\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.871485 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbfd688a-ee0f-495b-a915-f0345172d2cc-client-ca\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.972730 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbfd688a-ee0f-495b-a915-f0345172d2cc-client-ca\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.972886 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbfd688a-ee0f-495b-a915-f0345172d2cc-config\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.972956 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbfd688a-ee0f-495b-a915-f0345172d2cc-serving-cert\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.972995 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbfd688a-ee0f-495b-a915-f0345172d2cc-proxy-ca-bundles\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.973031 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8l4m\" (UniqueName: \"kubernetes.io/projected/cbfd688a-ee0f-495b-a915-f0345172d2cc-kube-api-access-r8l4m\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.974259 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbfd688a-ee0f-495b-a915-f0345172d2cc-client-ca\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.974492 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbfd688a-ee0f-495b-a915-f0345172d2cc-proxy-ca-bundles\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.977458 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbfd688a-ee0f-495b-a915-f0345172d2cc-config\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.982169 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbfd688a-ee0f-495b-a915-f0345172d2cc-serving-cert\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:35 crc kubenswrapper[4862]: I1128 11:27:35.991490 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8l4m\" (UniqueName: \"kubernetes.io/projected/cbfd688a-ee0f-495b-a915-f0345172d2cc-kube-api-access-r8l4m\") pod \"controller-manager-559bb546ff-8kkhz\" (UID: \"cbfd688a-ee0f-495b-a915-f0345172d2cc\") " pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:36 crc kubenswrapper[4862]: I1128 11:27:36.077374 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:36 crc kubenswrapper[4862]: I1128 11:27:36.497711 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-559bb546ff-8kkhz"] Nov 28 11:27:36 crc kubenswrapper[4862]: W1128 11:27:36.507914 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbfd688a_ee0f_495b_a915_f0345172d2cc.slice/crio-788048d214813de69c402e9391319cd8e1b4883642796edb63a2d208bf10bd20 WatchSource:0}: Error finding container 788048d214813de69c402e9391319cd8e1b4883642796edb63a2d208bf10bd20: Status 404 returned error can't find the container with id 788048d214813de69c402e9391319cd8e1b4883642796edb63a2d208bf10bd20 Nov 28 11:27:37 crc kubenswrapper[4862]: I1128 11:27:37.109268 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" event={"ID":"cbfd688a-ee0f-495b-a915-f0345172d2cc","Type":"ContainerStarted","Data":"ec7361491020a03e3af23cfa1f7d84d81acaade523043322a8f6bf18cb1631f6"} Nov 28 11:27:37 crc kubenswrapper[4862]: I1128 11:27:37.109616 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" event={"ID":"cbfd688a-ee0f-495b-a915-f0345172d2cc","Type":"ContainerStarted","Data":"788048d214813de69c402e9391319cd8e1b4883642796edb63a2d208bf10bd20"} Nov 28 11:27:37 crc kubenswrapper[4862]: I1128 11:27:37.109933 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:37 crc kubenswrapper[4862]: I1128 11:27:37.115118 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" Nov 28 11:27:37 crc kubenswrapper[4862]: I1128 11:27:37.131201 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-559bb546ff-8kkhz" podStartSLOduration=5.131175715 podStartE2EDuration="5.131175715s" podCreationTimestamp="2025-11-28 11:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:27:37.126882221 +0000 UTC m=+334.659396172" watchObservedRunningTime="2025-11-28 11:27:37.131175715 +0000 UTC m=+334.663689646" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.348520 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j5c55"] Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.349656 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j5c55" podUID="19d8710e-5f52-452e-8b13-854546736587" containerName="registry-server" containerID="cri-o://95abc85c195da8dad6596f05b3370a934f47a49cef7c6bdce53fa1ddacab0611" gracePeriod=30 Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.355302 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wbstd"] Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.355890 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wbstd" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerName="registry-server" containerID="cri-o://978c2dcc8a8182f2b01e80eeed8d829ad183f70dc89b51af05b1b3f05a6a4468" gracePeriod=30 Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.367579 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nbzb4"] Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.367836 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" podUID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerName="marketplace-operator" containerID="cri-o://2d7ab602e21fb4d1d4b1581a5674379862206b9bfe261952d3f640e5e4281f59" gracePeriod=30 Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.399945 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8jkt"] Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.400257 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l8jkt" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerName="registry-server" containerID="cri-o://059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714" gracePeriod=30 Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.412347 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w992j"] Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.412604 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w992j" podUID="02589a29-1121-48de-98a6-691d4f8c598f" containerName="registry-server" containerID="cri-o://a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0" gracePeriod=30 Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.419840 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q7zkb"] Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.442286 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: E1128 11:27:59.450397 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714 is running failed: container process not found" containerID="059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 11:27:59 crc kubenswrapper[4862]: E1128 11:27:59.451118 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714 is running failed: container process not found" containerID="059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 11:27:59 crc kubenswrapper[4862]: E1128 11:27:59.451571 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714 is running failed: container process not found" containerID="059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 11:27:59 crc kubenswrapper[4862]: E1128 11:27:59.451613 4862 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-l8jkt" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerName="registry-server" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.461169 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q7zkb"] Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.548950 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxsqd\" (UniqueName: \"kubernetes.io/projected/db7b8556-bd2d-4ed1-b3c2-984e207ae30b-kube-api-access-wxsqd\") pod \"marketplace-operator-79b997595-q7zkb\" (UID: \"db7b8556-bd2d-4ed1-b3c2-984e207ae30b\") " pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.549025 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db7b8556-bd2d-4ed1-b3c2-984e207ae30b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q7zkb\" (UID: \"db7b8556-bd2d-4ed1-b3c2-984e207ae30b\") " pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.549183 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db7b8556-bd2d-4ed1-b3c2-984e207ae30b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q7zkb\" (UID: \"db7b8556-bd2d-4ed1-b3c2-984e207ae30b\") " pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.650608 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxsqd\" (UniqueName: \"kubernetes.io/projected/db7b8556-bd2d-4ed1-b3c2-984e207ae30b-kube-api-access-wxsqd\") pod \"marketplace-operator-79b997595-q7zkb\" (UID: \"db7b8556-bd2d-4ed1-b3c2-984e207ae30b\") " pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.650675 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db7b8556-bd2d-4ed1-b3c2-984e207ae30b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q7zkb\" (UID: \"db7b8556-bd2d-4ed1-b3c2-984e207ae30b\") " pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.650717 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db7b8556-bd2d-4ed1-b3c2-984e207ae30b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q7zkb\" (UID: \"db7b8556-bd2d-4ed1-b3c2-984e207ae30b\") " pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.652263 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db7b8556-bd2d-4ed1-b3c2-984e207ae30b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q7zkb\" (UID: \"db7b8556-bd2d-4ed1-b3c2-984e207ae30b\") " pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.656572 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db7b8556-bd2d-4ed1-b3c2-984e207ae30b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q7zkb\" (UID: \"db7b8556-bd2d-4ed1-b3c2-984e207ae30b\") " pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.674854 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxsqd\" (UniqueName: \"kubernetes.io/projected/db7b8556-bd2d-4ed1-b3c2-984e207ae30b-kube-api-access-wxsqd\") pod \"marketplace-operator-79b997595-q7zkb\" (UID: \"db7b8556-bd2d-4ed1-b3c2-984e207ae30b\") " pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.801889 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:27:59 crc kubenswrapper[4862]: I1128 11:27:59.939407 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:27:59.985767 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.063377 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-catalog-content\") pod \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.063461 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-utilities\") pod \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.063490 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwbpj\" (UniqueName: \"kubernetes.io/projected/02589a29-1121-48de-98a6-691d4f8c598f-kube-api-access-kwbpj\") pod \"02589a29-1121-48de-98a6-691d4f8c598f\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.063527 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-catalog-content\") pod \"02589a29-1121-48de-98a6-691d4f8c598f\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.063545 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-utilities\") pod \"02589a29-1121-48de-98a6-691d4f8c598f\" (UID: \"02589a29-1121-48de-98a6-691d4f8c598f\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.063577 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7cmt\" (UniqueName: \"kubernetes.io/projected/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-kube-api-access-m7cmt\") pod \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\" (UID: \"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.064802 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-utilities" (OuterVolumeSpecName: "utilities") pod "9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" (UID: "9ded77a9-e0a3-410f-b118-8ab6e7be7bcf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.065364 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-utilities" (OuterVolumeSpecName: "utilities") pod "02589a29-1121-48de-98a6-691d4f8c598f" (UID: "02589a29-1121-48de-98a6-691d4f8c598f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.070724 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02589a29-1121-48de-98a6-691d4f8c598f-kube-api-access-kwbpj" (OuterVolumeSpecName: "kube-api-access-kwbpj") pod "02589a29-1121-48de-98a6-691d4f8c598f" (UID: "02589a29-1121-48de-98a6-691d4f8c598f"). InnerVolumeSpecName "kube-api-access-kwbpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.071030 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-kube-api-access-m7cmt" (OuterVolumeSpecName: "kube-api-access-m7cmt") pod "9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" (UID: "9ded77a9-e0a3-410f-b118-8ab6e7be7bcf"). InnerVolumeSpecName "kube-api-access-m7cmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.087006 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" (UID: "9ded77a9-e0a3-410f-b118-8ab6e7be7bcf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.165873 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02589a29-1121-48de-98a6-691d4f8c598f" (UID: "02589a29-1121-48de-98a6-691d4f8c598f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.166025 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7cmt\" (UniqueName: \"kubernetes.io/projected/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-kube-api-access-m7cmt\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.166059 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.166072 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.166166 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwbpj\" (UniqueName: \"kubernetes.io/projected/02589a29-1121-48de-98a6-691d4f8c598f-kube-api-access-kwbpj\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.166180 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.258975 4862 generic.go:334] "Generic (PLEG): container finished" podID="19d8710e-5f52-452e-8b13-854546736587" containerID="95abc85c195da8dad6596f05b3370a934f47a49cef7c6bdce53fa1ddacab0611" exitCode=0 Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.259047 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5c55" event={"ID":"19d8710e-5f52-452e-8b13-854546736587","Type":"ContainerDied","Data":"95abc85c195da8dad6596f05b3370a934f47a49cef7c6bdce53fa1ddacab0611"} Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.263214 4862 generic.go:334] "Generic (PLEG): container finished" podID="02589a29-1121-48de-98a6-691d4f8c598f" containerID="a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0" exitCode=0 Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.263280 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w992j" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.263286 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w992j" event={"ID":"02589a29-1121-48de-98a6-691d4f8c598f","Type":"ContainerDied","Data":"a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0"} Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.263313 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w992j" event={"ID":"02589a29-1121-48de-98a6-691d4f8c598f","Type":"ContainerDied","Data":"71cae4f273e0d20052b93ac8e4efd9927c7637721bab89e8ba1982c1f655186e"} Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.263330 4862 scope.go:117] "RemoveContainer" containerID="a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.266425 4862 generic.go:334] "Generic (PLEG): container finished" podID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerID="978c2dcc8a8182f2b01e80eeed8d829ad183f70dc89b51af05b1b3f05a6a4468" exitCode=0 Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.266528 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbstd" event={"ID":"2f359e91-8745-46c0-acad-6bc9795f04ba","Type":"ContainerDied","Data":"978c2dcc8a8182f2b01e80eeed8d829ad183f70dc89b51af05b1b3f05a6a4468"} Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.266876 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02589a29-1121-48de-98a6-691d4f8c598f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.269501 4862 generic.go:334] "Generic (PLEG): container finished" podID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerID="2d7ab602e21fb4d1d4b1581a5674379862206b9bfe261952d3f640e5e4281f59" exitCode=0 Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.269571 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" event={"ID":"3b577ecb-1896-4699-bd7e-33929f2cf9a3","Type":"ContainerDied","Data":"2d7ab602e21fb4d1d4b1581a5674379862206b9bfe261952d3f640e5e4281f59"} Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.275280 4862 generic.go:334] "Generic (PLEG): container finished" podID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerID="059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714" exitCode=0 Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.275313 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8jkt" event={"ID":"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf","Type":"ContainerDied","Data":"059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714"} Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.275334 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8jkt" event={"ID":"9ded77a9-e0a3-410f-b118-8ab6e7be7bcf","Type":"ContainerDied","Data":"d46a19c45f4f0b1f281eca375e9ad5642dd44e0dab93d230b246541dd744c2b6"} Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.275377 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l8jkt" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.297017 4862 scope.go:117] "RemoveContainer" containerID="34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.306452 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w992j"] Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.311351 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w992j"] Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.328775 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8jkt"] Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.334442 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8jkt"] Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.335815 4862 scope.go:117] "RemoveContainer" containerID="08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.373482 4862 scope.go:117] "RemoveContainer" containerID="a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0" Nov 28 11:28:00 crc kubenswrapper[4862]: E1128 11:28:00.375436 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0\": container with ID starting with a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0 not found: ID does not exist" containerID="a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.375718 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0"} err="failed to get container status \"a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0\": rpc error: code = NotFound desc = could not find container \"a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0\": container with ID starting with a1ef053ce5ceb59fa596f51d7a81fddb712d02861b1e88913e3bc96ab9c227e0 not found: ID does not exist" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.375755 4862 scope.go:117] "RemoveContainer" containerID="34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446" Nov 28 11:28:00 crc kubenswrapper[4862]: E1128 11:28:00.376127 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446\": container with ID starting with 34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446 not found: ID does not exist" containerID="34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.376162 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446"} err="failed to get container status \"34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446\": rpc error: code = NotFound desc = could not find container \"34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446\": container with ID starting with 34ed486d1dd08c471075534a5d3e4e25f193fcd25d3c9d916983e28dcf8ed446 not found: ID does not exist" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.376185 4862 scope.go:117] "RemoveContainer" containerID="08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480" Nov 28 11:28:00 crc kubenswrapper[4862]: E1128 11:28:00.376581 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480\": container with ID starting with 08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480 not found: ID does not exist" containerID="08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.376615 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480"} err="failed to get container status \"08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480\": rpc error: code = NotFound desc = could not find container \"08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480\": container with ID starting with 08e83a4842e0e398f7a978e0bfc71b6e745b60602f24af81f92b01a234408480 not found: ID does not exist" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.376635 4862 scope.go:117] "RemoveContainer" containerID="7ce732110dabdb015cb436adb7d4492085119d3ed7379ec98d25dff4366d4872" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.422149 4862 scope.go:117] "RemoveContainer" containerID="059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.441190 4862 scope.go:117] "RemoveContainer" containerID="3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.484065 4862 scope.go:117] "RemoveContainer" containerID="6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.525137 4862 scope.go:117] "RemoveContainer" containerID="059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714" Nov 28 11:28:00 crc kubenswrapper[4862]: E1128 11:28:00.525478 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714\": container with ID starting with 059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714 not found: ID does not exist" containerID="059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.525517 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714"} err="failed to get container status \"059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714\": rpc error: code = NotFound desc = could not find container \"059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714\": container with ID starting with 059842e1fd650629ce0966353d27c2fe6bbba50658879d7dc3f51f74105ab714 not found: ID does not exist" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.525545 4862 scope.go:117] "RemoveContainer" containerID="3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a" Nov 28 11:28:00 crc kubenswrapper[4862]: E1128 11:28:00.525765 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a\": container with ID starting with 3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a not found: ID does not exist" containerID="3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.525787 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a"} err="failed to get container status \"3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a\": rpc error: code = NotFound desc = could not find container \"3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a\": container with ID starting with 3233fcedc868cdb5a4869f1a0296b1e22f776015110607221c687f971abc8c8a not found: ID does not exist" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.525805 4862 scope.go:117] "RemoveContainer" containerID="6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06" Nov 28 11:28:00 crc kubenswrapper[4862]: E1128 11:28:00.526042 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06\": container with ID starting with 6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06 not found: ID does not exist" containerID="6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.526068 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06"} err="failed to get container status \"6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06\": rpc error: code = NotFound desc = could not find container \"6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06\": container with ID starting with 6f251210b033dac7ae9d3310d2c931b7b8beebcefeb332207799238666269e06 not found: ID does not exist" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.542804 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q7zkb"] Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.544916 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.641883 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.651562 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.671612 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-utilities\") pod \"2f359e91-8745-46c0-acad-6bc9795f04ba\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.671696 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfzcg\" (UniqueName: \"kubernetes.io/projected/2f359e91-8745-46c0-acad-6bc9795f04ba-kube-api-access-wfzcg\") pod \"2f359e91-8745-46c0-acad-6bc9795f04ba\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.671724 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-catalog-content\") pod \"2f359e91-8745-46c0-acad-6bc9795f04ba\" (UID: \"2f359e91-8745-46c0-acad-6bc9795f04ba\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.672697 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-utilities" (OuterVolumeSpecName: "utilities") pod "2f359e91-8745-46c0-acad-6bc9795f04ba" (UID: "2f359e91-8745-46c0-acad-6bc9795f04ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.680452 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f359e91-8745-46c0-acad-6bc9795f04ba-kube-api-access-wfzcg" (OuterVolumeSpecName: "kube-api-access-wfzcg") pod "2f359e91-8745-46c0-acad-6bc9795f04ba" (UID: "2f359e91-8745-46c0-acad-6bc9795f04ba"). InnerVolumeSpecName "kube-api-access-wfzcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.730035 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f359e91-8745-46c0-acad-6bc9795f04ba" (UID: "2f359e91-8745-46c0-acad-6bc9795f04ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.772991 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-utilities\") pod \"19d8710e-5f52-452e-8b13-854546736587\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.773038 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-trusted-ca\") pod \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.773074 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-operator-metrics\") pod \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.773134 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-catalog-content\") pod \"19d8710e-5f52-452e-8b13-854546736587\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.773164 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfklc\" (UniqueName: \"kubernetes.io/projected/3b577ecb-1896-4699-bd7e-33929f2cf9a3-kube-api-access-cfklc\") pod \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\" (UID: \"3b577ecb-1896-4699-bd7e-33929f2cf9a3\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.773185 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvgdr\" (UniqueName: \"kubernetes.io/projected/19d8710e-5f52-452e-8b13-854546736587-kube-api-access-tvgdr\") pod \"19d8710e-5f52-452e-8b13-854546736587\" (UID: \"19d8710e-5f52-452e-8b13-854546736587\") " Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.773434 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.773447 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfzcg\" (UniqueName: \"kubernetes.io/projected/2f359e91-8745-46c0-acad-6bc9795f04ba-kube-api-access-wfzcg\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.773459 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f359e91-8745-46c0-acad-6bc9795f04ba-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.774359 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-utilities" (OuterVolumeSpecName: "utilities") pod "19d8710e-5f52-452e-8b13-854546736587" (UID: "19d8710e-5f52-452e-8b13-854546736587"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.775970 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d8710e-5f52-452e-8b13-854546736587-kube-api-access-tvgdr" (OuterVolumeSpecName: "kube-api-access-tvgdr") pod "19d8710e-5f52-452e-8b13-854546736587" (UID: "19d8710e-5f52-452e-8b13-854546736587"). InnerVolumeSpecName "kube-api-access-tvgdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.776964 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "3b577ecb-1896-4699-bd7e-33929f2cf9a3" (UID: "3b577ecb-1896-4699-bd7e-33929f2cf9a3"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.778436 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b577ecb-1896-4699-bd7e-33929f2cf9a3-kube-api-access-cfklc" (OuterVolumeSpecName: "kube-api-access-cfklc") pod "3b577ecb-1896-4699-bd7e-33929f2cf9a3" (UID: "3b577ecb-1896-4699-bd7e-33929f2cf9a3"). InnerVolumeSpecName "kube-api-access-cfklc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.780224 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "3b577ecb-1896-4699-bd7e-33929f2cf9a3" (UID: "3b577ecb-1896-4699-bd7e-33929f2cf9a3"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.845021 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02589a29-1121-48de-98a6-691d4f8c598f" path="/var/lib/kubelet/pods/02589a29-1121-48de-98a6-691d4f8c598f/volumes" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.845855 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" path="/var/lib/kubelet/pods/9ded77a9-e0a3-410f-b118-8ab6e7be7bcf/volumes" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.849296 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19d8710e-5f52-452e-8b13-854546736587" (UID: "19d8710e-5f52-452e-8b13-854546736587"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.874661 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfklc\" (UniqueName: \"kubernetes.io/projected/3b577ecb-1896-4699-bd7e-33929f2cf9a3-kube-api-access-cfklc\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.874697 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvgdr\" (UniqueName: \"kubernetes.io/projected/19d8710e-5f52-452e-8b13-854546736587-kube-api-access-tvgdr\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.874711 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.874727 4862 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.874743 4862 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b577ecb-1896-4699-bd7e-33929f2cf9a3-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:00 crc kubenswrapper[4862]: I1128 11:28:00.874756 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d8710e-5f52-452e-8b13-854546736587-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.281394 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" event={"ID":"3b577ecb-1896-4699-bd7e-33929f2cf9a3","Type":"ContainerDied","Data":"f1869b57ee704bc529dc264795865c3f872df7083eb41a0df5814c02ae0f142a"} Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.281450 4862 scope.go:117] "RemoveContainer" containerID="2d7ab602e21fb4d1d4b1581a5674379862206b9bfe261952d3f640e5e4281f59" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.281412 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nbzb4" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.289905 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5c55" event={"ID":"19d8710e-5f52-452e-8b13-854546736587","Type":"ContainerDied","Data":"12204150939db276c9bca43802ce0e06722f2176e05bbea7c33bd81e7652836b"} Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.289985 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j5c55" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.296040 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbstd" event={"ID":"2f359e91-8745-46c0-acad-6bc9795f04ba","Type":"ContainerDied","Data":"7182ba62bae339490dc42abeea09fd665f32ccbe642f17b17a18991caba0fc75"} Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.296161 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbstd" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.299178 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" event={"ID":"db7b8556-bd2d-4ed1-b3c2-984e207ae30b","Type":"ContainerStarted","Data":"ff61b3c5227f8769baa739916c8f295e7d0e9d738a2ef344b9db971e6456e4d5"} Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.299257 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" event={"ID":"db7b8556-bd2d-4ed1-b3c2-984e207ae30b","Type":"ContainerStarted","Data":"5dda04c483a0ef9385d123ceea4cb30e3a2bcdf33ae7cde327660a0cc1cfdde8"} Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.299304 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.305350 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nbzb4"] Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.308384 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nbzb4"] Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.309073 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.311048 4862 scope.go:117] "RemoveContainer" containerID="95abc85c195da8dad6596f05b3370a934f47a49cef7c6bdce53fa1ddacab0611" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.338679 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-q7zkb" podStartSLOduration=2.338642706 podStartE2EDuration="2.338642706s" podCreationTimestamp="2025-11-28 11:27:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:28:01.332005813 +0000 UTC m=+358.864519784" watchObservedRunningTime="2025-11-28 11:28:01.338642706 +0000 UTC m=+358.871156677" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.339293 4862 scope.go:117] "RemoveContainer" containerID="ac8f2909e2d2f83d2a9587c9c051fdc44a2111f5ba3ce35134ef8ae5d1c19bb0" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.367784 4862 scope.go:117] "RemoveContainer" containerID="013c87cac0693d27468d902f9763c86a0d8bde97f800b0f450b5f7a445ba5c43" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.394511 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wbstd"] Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.399602 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wbstd"] Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.416654 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j5c55"] Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.416951 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j5c55"] Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.427622 4862 scope.go:117] "RemoveContainer" containerID="978c2dcc8a8182f2b01e80eeed8d829ad183f70dc89b51af05b1b3f05a6a4468" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.441526 4862 scope.go:117] "RemoveContainer" containerID="faa354060b696b1d143863519bd427d0390e78410605b8f4749c3d0780ffdf60" Nov 28 11:28:01 crc kubenswrapper[4862]: I1128 11:28:01.454741 4862 scope.go:117] "RemoveContainer" containerID="4c8ff714cb99c1551fabc9f5393064ba17c22e985ca7522803598711941cbcee" Nov 28 11:28:02 crc kubenswrapper[4862]: I1128 11:28:02.852613 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19d8710e-5f52-452e-8b13-854546736587" path="/var/lib/kubelet/pods/19d8710e-5f52-452e-8b13-854546736587/volumes" Nov 28 11:28:02 crc kubenswrapper[4862]: I1128 11:28:02.853405 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" path="/var/lib/kubelet/pods/2f359e91-8745-46c0-acad-6bc9795f04ba/volumes" Nov 28 11:28:02 crc kubenswrapper[4862]: I1128 11:28:02.854027 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" path="/var/lib/kubelet/pods/3b577ecb-1896-4699-bd7e-33929f2cf9a3/volumes" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.128636 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tk9wc"] Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129512 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d8710e-5f52-452e-8b13-854546736587" containerName="extract-content" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129536 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d8710e-5f52-452e-8b13-854546736587" containerName="extract-content" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129556 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d8710e-5f52-452e-8b13-854546736587" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129568 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d8710e-5f52-452e-8b13-854546736587" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129588 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02589a29-1121-48de-98a6-691d4f8c598f" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129600 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="02589a29-1121-48de-98a6-691d4f8c598f" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129616 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerName="extract-utilities" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129628 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerName="extract-utilities" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129644 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerName="extract-content" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129658 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerName="extract-content" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129680 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02589a29-1121-48de-98a6-691d4f8c598f" containerName="extract-utilities" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129692 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="02589a29-1121-48de-98a6-691d4f8c598f" containerName="extract-utilities" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129706 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129718 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129736 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerName="extract-utilities" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129747 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerName="extract-utilities" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129761 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02589a29-1121-48de-98a6-691d4f8c598f" containerName="extract-content" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129772 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="02589a29-1121-48de-98a6-691d4f8c598f" containerName="extract-content" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129786 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d8710e-5f52-452e-8b13-854546736587" containerName="extract-utilities" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129797 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d8710e-5f52-452e-8b13-854546736587" containerName="extract-utilities" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129818 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129829 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129847 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerName="marketplace-operator" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129858 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerName="marketplace-operator" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129881 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerName="marketplace-operator" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129893 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerName="marketplace-operator" Nov 28 11:28:05 crc kubenswrapper[4862]: E1128 11:28:05.129914 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerName="extract-content" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.129926 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerName="extract-content" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.130198 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="02589a29-1121-48de-98a6-691d4f8c598f" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.130224 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f359e91-8745-46c0-acad-6bc9795f04ba" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.130243 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d8710e-5f52-452e-8b13-854546736587" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.130255 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerName="marketplace-operator" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.130278 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ded77a9-e0a3-410f-b118-8ab6e7be7bcf" containerName="registry-server" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.130295 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b577ecb-1896-4699-bd7e-33929f2cf9a3" containerName="marketplace-operator" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.130852 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.141337 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tk9wc"] Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.235857 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2aa7504-65f3-45b9-a15c-f2603822120a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.235919 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpqd8\" (UniqueName: \"kubernetes.io/projected/e2aa7504-65f3-45b9-a15c-f2603822120a-kube-api-access-cpqd8\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.235947 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2aa7504-65f3-45b9-a15c-f2603822120a-trusted-ca\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.235969 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2aa7504-65f3-45b9-a15c-f2603822120a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.236154 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.236246 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2aa7504-65f3-45b9-a15c-f2603822120a-registry-certificates\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.236291 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2aa7504-65f3-45b9-a15c-f2603822120a-bound-sa-token\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.236524 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2aa7504-65f3-45b9-a15c-f2603822120a-registry-tls\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.302718 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.337676 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2aa7504-65f3-45b9-a15c-f2603822120a-registry-tls\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.337718 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2aa7504-65f3-45b9-a15c-f2603822120a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.337740 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpqd8\" (UniqueName: \"kubernetes.io/projected/e2aa7504-65f3-45b9-a15c-f2603822120a-kube-api-access-cpqd8\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.337765 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2aa7504-65f3-45b9-a15c-f2603822120a-trusted-ca\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.337784 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2aa7504-65f3-45b9-a15c-f2603822120a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.337813 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2aa7504-65f3-45b9-a15c-f2603822120a-registry-certificates\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.337834 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2aa7504-65f3-45b9-a15c-f2603822120a-bound-sa-token\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.338982 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2aa7504-65f3-45b9-a15c-f2603822120a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.340023 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2aa7504-65f3-45b9-a15c-f2603822120a-trusted-ca\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.340160 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2aa7504-65f3-45b9-a15c-f2603822120a-registry-certificates\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.345162 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2aa7504-65f3-45b9-a15c-f2603822120a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.352960 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpqd8\" (UniqueName: \"kubernetes.io/projected/e2aa7504-65f3-45b9-a15c-f2603822120a-kube-api-access-cpqd8\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.357767 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2aa7504-65f3-45b9-a15c-f2603822120a-registry-tls\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.372675 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2aa7504-65f3-45b9-a15c-f2603822120a-bound-sa-token\") pod \"image-registry-66df7c8f76-tk9wc\" (UID: \"e2aa7504-65f3-45b9-a15c-f2603822120a\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.447906 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:05 crc kubenswrapper[4862]: I1128 11:28:05.893461 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tk9wc"] Nov 28 11:28:06 crc kubenswrapper[4862]: I1128 11:28:06.338267 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" event={"ID":"e2aa7504-65f3-45b9-a15c-f2603822120a","Type":"ContainerStarted","Data":"00d7c95d5fd1b5a17bb542473211007f220d311e7467950c2144ea7ca49c66e3"} Nov 28 11:28:06 crc kubenswrapper[4862]: I1128 11:28:06.338671 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:06 crc kubenswrapper[4862]: I1128 11:28:06.338695 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" event={"ID":"e2aa7504-65f3-45b9-a15c-f2603822120a","Type":"ContainerStarted","Data":"963a96b6ef3f318c493883ad679634e258ea8404852ab353c8a2a01039cc910e"} Nov 28 11:28:06 crc kubenswrapper[4862]: I1128 11:28:06.361110 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" podStartSLOduration=1.361078852 podStartE2EDuration="1.361078852s" podCreationTimestamp="2025-11-28 11:28:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:28:06.356987273 +0000 UTC m=+363.889501214" watchObservedRunningTime="2025-11-28 11:28:06.361078852 +0000 UTC m=+363.893592793" Nov 28 11:28:08 crc kubenswrapper[4862]: I1128 11:28:08.291594 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:28:08 crc kubenswrapper[4862]: I1128 11:28:08.292017 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.382363 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2sz82"] Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.384639 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.387948 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.398826 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2sz82"] Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.421957 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93cd8790-a274-4fdb-8c48-028231776561-utilities\") pod \"certified-operators-2sz82\" (UID: \"93cd8790-a274-4fdb-8c48-028231776561\") " pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.422035 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93cd8790-a274-4fdb-8c48-028231776561-catalog-content\") pod \"certified-operators-2sz82\" (UID: \"93cd8790-a274-4fdb-8c48-028231776561\") " pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.422118 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thzfk\" (UniqueName: \"kubernetes.io/projected/93cd8790-a274-4fdb-8c48-028231776561-kube-api-access-thzfk\") pod \"certified-operators-2sz82\" (UID: \"93cd8790-a274-4fdb-8c48-028231776561\") " pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.523863 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thzfk\" (UniqueName: \"kubernetes.io/projected/93cd8790-a274-4fdb-8c48-028231776561-kube-api-access-thzfk\") pod \"certified-operators-2sz82\" (UID: \"93cd8790-a274-4fdb-8c48-028231776561\") " pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.524018 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93cd8790-a274-4fdb-8c48-028231776561-utilities\") pod \"certified-operators-2sz82\" (UID: \"93cd8790-a274-4fdb-8c48-028231776561\") " pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.524114 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93cd8790-a274-4fdb-8c48-028231776561-catalog-content\") pod \"certified-operators-2sz82\" (UID: \"93cd8790-a274-4fdb-8c48-028231776561\") " pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.524896 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93cd8790-a274-4fdb-8c48-028231776561-utilities\") pod \"certified-operators-2sz82\" (UID: \"93cd8790-a274-4fdb-8c48-028231776561\") " pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.524990 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93cd8790-a274-4fdb-8c48-028231776561-catalog-content\") pod \"certified-operators-2sz82\" (UID: \"93cd8790-a274-4fdb-8c48-028231776561\") " pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.547642 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thzfk\" (UniqueName: \"kubernetes.io/projected/93cd8790-a274-4fdb-8c48-028231776561-kube-api-access-thzfk\") pod \"certified-operators-2sz82\" (UID: \"93cd8790-a274-4fdb-8c48-028231776561\") " pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.590645 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q66nw"] Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.593165 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.596469 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.601564 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q66nw"] Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.708732 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.726805 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a08c877-6f2a-4c68-8764-f9cb39e16387-catalog-content\") pod \"community-operators-q66nw\" (UID: \"6a08c877-6f2a-4c68-8764-f9cb39e16387\") " pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.726850 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a08c877-6f2a-4c68-8764-f9cb39e16387-utilities\") pod \"community-operators-q66nw\" (UID: \"6a08c877-6f2a-4c68-8764-f9cb39e16387\") " pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.726885 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5mkf\" (UniqueName: \"kubernetes.io/projected/6a08c877-6f2a-4c68-8764-f9cb39e16387-kube-api-access-t5mkf\") pod \"community-operators-q66nw\" (UID: \"6a08c877-6f2a-4c68-8764-f9cb39e16387\") " pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.829712 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a08c877-6f2a-4c68-8764-f9cb39e16387-catalog-content\") pod \"community-operators-q66nw\" (UID: \"6a08c877-6f2a-4c68-8764-f9cb39e16387\") " pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.830210 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a08c877-6f2a-4c68-8764-f9cb39e16387-utilities\") pod \"community-operators-q66nw\" (UID: \"6a08c877-6f2a-4c68-8764-f9cb39e16387\") " pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.830250 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5mkf\" (UniqueName: \"kubernetes.io/projected/6a08c877-6f2a-4c68-8764-f9cb39e16387-kube-api-access-t5mkf\") pod \"community-operators-q66nw\" (UID: \"6a08c877-6f2a-4c68-8764-f9cb39e16387\") " pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.831432 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a08c877-6f2a-4c68-8764-f9cb39e16387-catalog-content\") pod \"community-operators-q66nw\" (UID: \"6a08c877-6f2a-4c68-8764-f9cb39e16387\") " pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.831511 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a08c877-6f2a-4c68-8764-f9cb39e16387-utilities\") pod \"community-operators-q66nw\" (UID: \"6a08c877-6f2a-4c68-8764-f9cb39e16387\") " pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.859288 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5mkf\" (UniqueName: \"kubernetes.io/projected/6a08c877-6f2a-4c68-8764-f9cb39e16387-kube-api-access-t5mkf\") pod \"community-operators-q66nw\" (UID: \"6a08c877-6f2a-4c68-8764-f9cb39e16387\") " pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:10 crc kubenswrapper[4862]: I1128 11:28:10.926016 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:11 crc kubenswrapper[4862]: I1128 11:28:11.164987 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2sz82"] Nov 28 11:28:11 crc kubenswrapper[4862]: I1128 11:28:11.386665 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sz82" event={"ID":"93cd8790-a274-4fdb-8c48-028231776561","Type":"ContainerStarted","Data":"0886947e551cc62978feda57e57b26f88d11807f26fa2d0e249b500f544a784c"} Nov 28 11:28:11 crc kubenswrapper[4862]: I1128 11:28:11.387250 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sz82" event={"ID":"93cd8790-a274-4fdb-8c48-028231776561","Type":"ContainerStarted","Data":"4d751a66f01daac18f80632d1a8878678f72862452b76548a4e3041d97c4e6e9"} Nov 28 11:28:11 crc kubenswrapper[4862]: I1128 11:28:11.423283 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q66nw"] Nov 28 11:28:11 crc kubenswrapper[4862]: W1128 11:28:11.483360 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a08c877_6f2a_4c68_8764_f9cb39e16387.slice/crio-7bc9914eee137727178f8d6cec3d663a04c194d9653afb266e5c026b21c71fa1 WatchSource:0}: Error finding container 7bc9914eee137727178f8d6cec3d663a04c194d9653afb266e5c026b21c71fa1: Status 404 returned error can't find the container with id 7bc9914eee137727178f8d6cec3d663a04c194d9653afb266e5c026b21c71fa1 Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.395964 4862 generic.go:334] "Generic (PLEG): container finished" podID="6a08c877-6f2a-4c68-8764-f9cb39e16387" containerID="43cd826cb13c841a9d09f0beab743fafbb6598cd6ca6cc30a19e11d7b535eba7" exitCode=0 Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.396035 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q66nw" event={"ID":"6a08c877-6f2a-4c68-8764-f9cb39e16387","Type":"ContainerDied","Data":"43cd826cb13c841a9d09f0beab743fafbb6598cd6ca6cc30a19e11d7b535eba7"} Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.396462 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q66nw" event={"ID":"6a08c877-6f2a-4c68-8764-f9cb39e16387","Type":"ContainerStarted","Data":"7bc9914eee137727178f8d6cec3d663a04c194d9653afb266e5c026b21c71fa1"} Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.398071 4862 generic.go:334] "Generic (PLEG): container finished" podID="93cd8790-a274-4fdb-8c48-028231776561" containerID="0886947e551cc62978feda57e57b26f88d11807f26fa2d0e249b500f544a784c" exitCode=0 Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.398121 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sz82" event={"ID":"93cd8790-a274-4fdb-8c48-028231776561","Type":"ContainerDied","Data":"0886947e551cc62978feda57e57b26f88d11807f26fa2d0e249b500f544a784c"} Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.789775 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f8552"] Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.792049 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.800064 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.809267 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8552"] Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.887556 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67725ff2-31c1-49ea-895e-36734cdab9ef-catalog-content\") pod \"redhat-marketplace-f8552\" (UID: \"67725ff2-31c1-49ea-895e-36734cdab9ef\") " pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.887607 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j286\" (UniqueName: \"kubernetes.io/projected/67725ff2-31c1-49ea-895e-36734cdab9ef-kube-api-access-4j286\") pod \"redhat-marketplace-f8552\" (UID: \"67725ff2-31c1-49ea-895e-36734cdab9ef\") " pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.888230 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67725ff2-31c1-49ea-895e-36734cdab9ef-utilities\") pod \"redhat-marketplace-f8552\" (UID: \"67725ff2-31c1-49ea-895e-36734cdab9ef\") " pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.981037 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dwqgj"] Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.982775 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.986970 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.989651 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j286\" (UniqueName: \"kubernetes.io/projected/67725ff2-31c1-49ea-895e-36734cdab9ef-kube-api-access-4j286\") pod \"redhat-marketplace-f8552\" (UID: \"67725ff2-31c1-49ea-895e-36734cdab9ef\") " pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.989825 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67725ff2-31c1-49ea-895e-36734cdab9ef-utilities\") pod \"redhat-marketplace-f8552\" (UID: \"67725ff2-31c1-49ea-895e-36734cdab9ef\") " pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.989890 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67725ff2-31c1-49ea-895e-36734cdab9ef-catalog-content\") pod \"redhat-marketplace-f8552\" (UID: \"67725ff2-31c1-49ea-895e-36734cdab9ef\") " pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.990628 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67725ff2-31c1-49ea-895e-36734cdab9ef-catalog-content\") pod \"redhat-marketplace-f8552\" (UID: \"67725ff2-31c1-49ea-895e-36734cdab9ef\") " pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.992286 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67725ff2-31c1-49ea-895e-36734cdab9ef-utilities\") pod \"redhat-marketplace-f8552\" (UID: \"67725ff2-31c1-49ea-895e-36734cdab9ef\") " pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:12 crc kubenswrapper[4862]: I1128 11:28:12.999719 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dwqgj"] Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.022885 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j286\" (UniqueName: \"kubernetes.io/projected/67725ff2-31c1-49ea-895e-36734cdab9ef-kube-api-access-4j286\") pod \"redhat-marketplace-f8552\" (UID: \"67725ff2-31c1-49ea-895e-36734cdab9ef\") " pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.092377 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de97f27c-53ed-42fa-ba3f-4fd817a2e32f-utilities\") pod \"redhat-operators-dwqgj\" (UID: \"de97f27c-53ed-42fa-ba3f-4fd817a2e32f\") " pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.092438 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de97f27c-53ed-42fa-ba3f-4fd817a2e32f-catalog-content\") pod \"redhat-operators-dwqgj\" (UID: \"de97f27c-53ed-42fa-ba3f-4fd817a2e32f\") " pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.092574 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtl7b\" (UniqueName: \"kubernetes.io/projected/de97f27c-53ed-42fa-ba3f-4fd817a2e32f-kube-api-access-wtl7b\") pod \"redhat-operators-dwqgj\" (UID: \"de97f27c-53ed-42fa-ba3f-4fd817a2e32f\") " pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.157477 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.194274 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de97f27c-53ed-42fa-ba3f-4fd817a2e32f-utilities\") pod \"redhat-operators-dwqgj\" (UID: \"de97f27c-53ed-42fa-ba3f-4fd817a2e32f\") " pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.194322 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de97f27c-53ed-42fa-ba3f-4fd817a2e32f-catalog-content\") pod \"redhat-operators-dwqgj\" (UID: \"de97f27c-53ed-42fa-ba3f-4fd817a2e32f\") " pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.194397 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtl7b\" (UniqueName: \"kubernetes.io/projected/de97f27c-53ed-42fa-ba3f-4fd817a2e32f-kube-api-access-wtl7b\") pod \"redhat-operators-dwqgj\" (UID: \"de97f27c-53ed-42fa-ba3f-4fd817a2e32f\") " pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.195967 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de97f27c-53ed-42fa-ba3f-4fd817a2e32f-catalog-content\") pod \"redhat-operators-dwqgj\" (UID: \"de97f27c-53ed-42fa-ba3f-4fd817a2e32f\") " pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.196216 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de97f27c-53ed-42fa-ba3f-4fd817a2e32f-utilities\") pod \"redhat-operators-dwqgj\" (UID: \"de97f27c-53ed-42fa-ba3f-4fd817a2e32f\") " pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.218984 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtl7b\" (UniqueName: \"kubernetes.io/projected/de97f27c-53ed-42fa-ba3f-4fd817a2e32f-kube-api-access-wtl7b\") pod \"redhat-operators-dwqgj\" (UID: \"de97f27c-53ed-42fa-ba3f-4fd817a2e32f\") " pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.315860 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.888487 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8552"] Nov 28 11:28:13 crc kubenswrapper[4862]: I1128 11:28:13.975733 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dwqgj"] Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.412363 4862 generic.go:334] "Generic (PLEG): container finished" podID="67725ff2-31c1-49ea-895e-36734cdab9ef" containerID="1f1077fff180931c64343064b99ac5cce297c4ebe4eb8a4c6d2e47b04c275068" exitCode=0 Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.412473 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8552" event={"ID":"67725ff2-31c1-49ea-895e-36734cdab9ef","Type":"ContainerDied","Data":"1f1077fff180931c64343064b99ac5cce297c4ebe4eb8a4c6d2e47b04c275068"} Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.412531 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8552" event={"ID":"67725ff2-31c1-49ea-895e-36734cdab9ef","Type":"ContainerStarted","Data":"482b768416b066167b8ad6a14b0e28f1aa06b6fe1138a540aeca561cf86ba96a"} Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.417790 4862 generic.go:334] "Generic (PLEG): container finished" podID="93cd8790-a274-4fdb-8c48-028231776561" containerID="9648fb64ea59f257a602cb5addf69ca9246a2d42bab26386fe2af144651411e3" exitCode=0 Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.417852 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sz82" event={"ID":"93cd8790-a274-4fdb-8c48-028231776561","Type":"ContainerDied","Data":"9648fb64ea59f257a602cb5addf69ca9246a2d42bab26386fe2af144651411e3"} Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.424848 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwqgj" event={"ID":"de97f27c-53ed-42fa-ba3f-4fd817a2e32f","Type":"ContainerDied","Data":"f6abc7dc237b5eaab69e551565b9c36b8dd89ac9e11c154c2ef74bf2457ef01f"} Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.425046 4862 generic.go:334] "Generic (PLEG): container finished" podID="de97f27c-53ed-42fa-ba3f-4fd817a2e32f" containerID="f6abc7dc237b5eaab69e551565b9c36b8dd89ac9e11c154c2ef74bf2457ef01f" exitCode=0 Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.425812 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwqgj" event={"ID":"de97f27c-53ed-42fa-ba3f-4fd817a2e32f","Type":"ContainerStarted","Data":"89206493a56015167984199bb36480d6346b9633fbb4503e783576c5a417ba8c"} Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.439006 4862 generic.go:334] "Generic (PLEG): container finished" podID="6a08c877-6f2a-4c68-8764-f9cb39e16387" containerID="c73e40b81763af2e26a3f9dab32db95f12d8bde7051065d873a57b0a266b0fbc" exitCode=0 Nov 28 11:28:14 crc kubenswrapper[4862]: I1128 11:28:14.439061 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q66nw" event={"ID":"6a08c877-6f2a-4c68-8764-f9cb39e16387","Type":"ContainerDied","Data":"c73e40b81763af2e26a3f9dab32db95f12d8bde7051065d873a57b0a266b0fbc"} Nov 28 11:28:15 crc kubenswrapper[4862]: I1128 11:28:15.448831 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q66nw" event={"ID":"6a08c877-6f2a-4c68-8764-f9cb39e16387","Type":"ContainerStarted","Data":"747a27b25916f2ff6aa1bcdebcbfd39703318d8daca63a42947ead699b313a0b"} Nov 28 11:28:15 crc kubenswrapper[4862]: I1128 11:28:15.488000 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q66nw" podStartSLOduration=2.727136894 podStartE2EDuration="5.48797109s" podCreationTimestamp="2025-11-28 11:28:10 +0000 UTC" firstStartedPulling="2025-11-28 11:28:12.399613605 +0000 UTC m=+369.932127526" lastFinishedPulling="2025-11-28 11:28:15.160447801 +0000 UTC m=+372.692961722" observedRunningTime="2025-11-28 11:28:15.480376549 +0000 UTC m=+373.012890470" watchObservedRunningTime="2025-11-28 11:28:15.48797109 +0000 UTC m=+373.020485011" Nov 28 11:28:16 crc kubenswrapper[4862]: I1128 11:28:16.457735 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sz82" event={"ID":"93cd8790-a274-4fdb-8c48-028231776561","Type":"ContainerStarted","Data":"3bbaaae1eb21912b7c3b86eb8fcecb8eeaa947dc55904420c96e85b508b368ed"} Nov 28 11:28:16 crc kubenswrapper[4862]: I1128 11:28:16.461846 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwqgj" event={"ID":"de97f27c-53ed-42fa-ba3f-4fd817a2e32f","Type":"ContainerStarted","Data":"b0f7feded06a0a5443d1a54fd617ce22ae9d330bf14a34f5e1786dfe56a9dfc6"} Nov 28 11:28:16 crc kubenswrapper[4862]: I1128 11:28:16.469343 4862 generic.go:334] "Generic (PLEG): container finished" podID="67725ff2-31c1-49ea-895e-36734cdab9ef" containerID="d5463be0a42c9e44e515207f093efe9f741907c4f8c48cf893ef42522d4c75f5" exitCode=0 Nov 28 11:28:16 crc kubenswrapper[4862]: I1128 11:28:16.470581 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8552" event={"ID":"67725ff2-31c1-49ea-895e-36734cdab9ef","Type":"ContainerDied","Data":"d5463be0a42c9e44e515207f093efe9f741907c4f8c48cf893ef42522d4c75f5"} Nov 28 11:28:16 crc kubenswrapper[4862]: I1128 11:28:16.530001 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2sz82" podStartSLOduration=3.079501822 podStartE2EDuration="6.529975899s" podCreationTimestamp="2025-11-28 11:28:10 +0000 UTC" firstStartedPulling="2025-11-28 11:28:12.399367147 +0000 UTC m=+369.931881078" lastFinishedPulling="2025-11-28 11:28:15.849841224 +0000 UTC m=+373.382355155" observedRunningTime="2025-11-28 11:28:16.525200641 +0000 UTC m=+374.057714592" watchObservedRunningTime="2025-11-28 11:28:16.529975899 +0000 UTC m=+374.062489850" Nov 28 11:28:17 crc kubenswrapper[4862]: I1128 11:28:17.479556 4862 generic.go:334] "Generic (PLEG): container finished" podID="de97f27c-53ed-42fa-ba3f-4fd817a2e32f" containerID="b0f7feded06a0a5443d1a54fd617ce22ae9d330bf14a34f5e1786dfe56a9dfc6" exitCode=0 Nov 28 11:28:17 crc kubenswrapper[4862]: I1128 11:28:17.481579 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwqgj" event={"ID":"de97f27c-53ed-42fa-ba3f-4fd817a2e32f","Type":"ContainerDied","Data":"b0f7feded06a0a5443d1a54fd617ce22ae9d330bf14a34f5e1786dfe56a9dfc6"} Nov 28 11:28:18 crc kubenswrapper[4862]: I1128 11:28:18.490029 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwqgj" event={"ID":"de97f27c-53ed-42fa-ba3f-4fd817a2e32f","Type":"ContainerStarted","Data":"ff58d0c04cc5132e0d4d489cf3723e39f9654d31f35b8d48189a2ea031318c42"} Nov 28 11:28:18 crc kubenswrapper[4862]: I1128 11:28:18.492490 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8552" event={"ID":"67725ff2-31c1-49ea-895e-36734cdab9ef","Type":"ContainerStarted","Data":"a4090d8d733fdc97a8c0656ce1b50e2d58f1f2d0d7bf3f6623007b0c94fefec7"} Nov 28 11:28:18 crc kubenswrapper[4862]: I1128 11:28:18.516661 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dwqgj" podStartSLOduration=2.905642563 podStartE2EDuration="6.516639675s" podCreationTimestamp="2025-11-28 11:28:12 +0000 UTC" firstStartedPulling="2025-11-28 11:28:14.426888208 +0000 UTC m=+371.959402129" lastFinishedPulling="2025-11-28 11:28:18.03788528 +0000 UTC m=+375.570399241" observedRunningTime="2025-11-28 11:28:18.51614368 +0000 UTC m=+376.048657641" watchObservedRunningTime="2025-11-28 11:28:18.516639675 +0000 UTC m=+376.049153596" Nov 28 11:28:18 crc kubenswrapper[4862]: I1128 11:28:18.545303 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f8552" podStartSLOduration=3.525469218 podStartE2EDuration="6.545276755s" podCreationTimestamp="2025-11-28 11:28:12 +0000 UTC" firstStartedPulling="2025-11-28 11:28:14.414109177 +0000 UTC m=+371.946623098" lastFinishedPulling="2025-11-28 11:28:17.433916714 +0000 UTC m=+374.966430635" observedRunningTime="2025-11-28 11:28:18.538977383 +0000 UTC m=+376.071491304" watchObservedRunningTime="2025-11-28 11:28:18.545276755 +0000 UTC m=+376.077790716" Nov 28 11:28:20 crc kubenswrapper[4862]: I1128 11:28:20.709274 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:20 crc kubenswrapper[4862]: I1128 11:28:20.709350 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:20 crc kubenswrapper[4862]: I1128 11:28:20.779157 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:20 crc kubenswrapper[4862]: I1128 11:28:20.926654 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:20 crc kubenswrapper[4862]: I1128 11:28:20.926758 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:20 crc kubenswrapper[4862]: I1128 11:28:20.989022 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:21 crc kubenswrapper[4862]: I1128 11:28:21.562152 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q66nw" Nov 28 11:28:21 crc kubenswrapper[4862]: I1128 11:28:21.575749 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2sz82" Nov 28 11:28:23 crc kubenswrapper[4862]: I1128 11:28:23.158789 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:23 crc kubenswrapper[4862]: I1128 11:28:23.159078 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:23 crc kubenswrapper[4862]: I1128 11:28:23.199891 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:23 crc kubenswrapper[4862]: I1128 11:28:23.317176 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:23 crc kubenswrapper[4862]: I1128 11:28:23.317248 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:23 crc kubenswrapper[4862]: I1128 11:28:23.571553 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f8552" Nov 28 11:28:24 crc kubenswrapper[4862]: I1128 11:28:24.360203 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dwqgj" podUID="de97f27c-53ed-42fa-ba3f-4fd817a2e32f" containerName="registry-server" probeResult="failure" output=< Nov 28 11:28:24 crc kubenswrapper[4862]: timeout: failed to connect service ":50051" within 1s Nov 28 11:28:24 crc kubenswrapper[4862]: > Nov 28 11:28:25 crc kubenswrapper[4862]: I1128 11:28:25.455129 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-tk9wc" Nov 28 11:28:25 crc kubenswrapper[4862]: I1128 11:28:25.565185 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hw6dt"] Nov 28 11:28:33 crc kubenswrapper[4862]: I1128 11:28:33.382954 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:33 crc kubenswrapper[4862]: I1128 11:28:33.461771 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dwqgj" Nov 28 11:28:38 crc kubenswrapper[4862]: I1128 11:28:38.293158 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:28:38 crc kubenswrapper[4862]: I1128 11:28:38.293794 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:28:50 crc kubenswrapper[4862]: I1128 11:28:50.620780 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" podUID="6e143937-4cf3-4fa0-aa09-381d8bf99f16" containerName="registry" containerID="cri-o://4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7" gracePeriod=30 Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.137488 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.246211 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-certificates\") pod \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.246299 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6e143937-4cf3-4fa0-aa09-381d8bf99f16-installation-pull-secrets\") pod \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.246591 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.246635 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-tls\") pod \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.246709 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-bound-sa-token\") pod \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.246780 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7plxg\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-kube-api-access-7plxg\") pod \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.246851 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-trusted-ca\") pod \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.246920 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6e143937-4cf3-4fa0-aa09-381d8bf99f16-ca-trust-extracted\") pod \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\" (UID: \"6e143937-4cf3-4fa0-aa09-381d8bf99f16\") " Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.248412 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "6e143937-4cf3-4fa0-aa09-381d8bf99f16" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.248454 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "6e143937-4cf3-4fa0-aa09-381d8bf99f16" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.256257 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "6e143937-4cf3-4fa0-aa09-381d8bf99f16" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.257877 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e143937-4cf3-4fa0-aa09-381d8bf99f16-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "6e143937-4cf3-4fa0-aa09-381d8bf99f16" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.258189 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-kube-api-access-7plxg" (OuterVolumeSpecName: "kube-api-access-7plxg") pod "6e143937-4cf3-4fa0-aa09-381d8bf99f16" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16"). InnerVolumeSpecName "kube-api-access-7plxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.258531 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "6e143937-4cf3-4fa0-aa09-381d8bf99f16" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.265247 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "6e143937-4cf3-4fa0-aa09-381d8bf99f16" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.288314 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e143937-4cf3-4fa0-aa09-381d8bf99f16-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "6e143937-4cf3-4fa0-aa09-381d8bf99f16" (UID: "6e143937-4cf3-4fa0-aa09-381d8bf99f16"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.348427 4862 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.348807 4862 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.348956 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7plxg\" (UniqueName: \"kubernetes.io/projected/6e143937-4cf3-4fa0-aa09-381d8bf99f16-kube-api-access-7plxg\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.349140 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.349276 4862 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6e143937-4cf3-4fa0-aa09-381d8bf99f16-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.349435 4862 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6e143937-4cf3-4fa0-aa09-381d8bf99f16-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.349585 4862 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6e143937-4cf3-4fa0-aa09-381d8bf99f16-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.731706 4862 generic.go:334] "Generic (PLEG): container finished" podID="6e143937-4cf3-4fa0-aa09-381d8bf99f16" containerID="4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7" exitCode=0 Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.731835 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.731847 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" event={"ID":"6e143937-4cf3-4fa0-aa09-381d8bf99f16","Type":"ContainerDied","Data":"4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7"} Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.732474 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hw6dt" event={"ID":"6e143937-4cf3-4fa0-aa09-381d8bf99f16","Type":"ContainerDied","Data":"f27a46368de3a5a6898ace1485ade3e48dbe69ecb1ded058eab1519f876937b4"} Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.732512 4862 scope.go:117] "RemoveContainer" containerID="4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.766802 4862 scope.go:117] "RemoveContainer" containerID="4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7" Nov 28 11:28:51 crc kubenswrapper[4862]: E1128 11:28:51.768898 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7\": container with ID starting with 4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7 not found: ID does not exist" containerID="4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.768971 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7"} err="failed to get container status \"4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7\": rpc error: code = NotFound desc = could not find container \"4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7\": container with ID starting with 4aefdb4891add81576ee9f6bb2bc6a60638de1809ed8dfab9f2c1dd214cd1ac7 not found: ID does not exist" Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.806381 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hw6dt"] Nov 28 11:28:51 crc kubenswrapper[4862]: I1128 11:28:51.810530 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hw6dt"] Nov 28 11:28:52 crc kubenswrapper[4862]: I1128 11:28:52.853004 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e143937-4cf3-4fa0-aa09-381d8bf99f16" path="/var/lib/kubelet/pods/6e143937-4cf3-4fa0-aa09-381d8bf99f16/volumes" Nov 28 11:29:08 crc kubenswrapper[4862]: I1128 11:29:08.292006 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:29:08 crc kubenswrapper[4862]: I1128 11:29:08.292942 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:29:08 crc kubenswrapper[4862]: I1128 11:29:08.293046 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:29:08 crc kubenswrapper[4862]: I1128 11:29:08.294683 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b510eb115f59f27520a8dc8f9d8b64d0f26aef3c7c9a1efef94445ccf2f2eb72"} pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:29:08 crc kubenswrapper[4862]: I1128 11:29:08.294797 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" containerID="cri-o://b510eb115f59f27520a8dc8f9d8b64d0f26aef3c7c9a1efef94445ccf2f2eb72" gracePeriod=600 Nov 28 11:29:08 crc kubenswrapper[4862]: I1128 11:29:08.852890 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6845439-1a4c-4f28-9ba4-29149d545995" containerID="b510eb115f59f27520a8dc8f9d8b64d0f26aef3c7c9a1efef94445ccf2f2eb72" exitCode=0 Nov 28 11:29:08 crc kubenswrapper[4862]: I1128 11:29:08.852946 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerDied","Data":"b510eb115f59f27520a8dc8f9d8b64d0f26aef3c7c9a1efef94445ccf2f2eb72"} Nov 28 11:29:08 crc kubenswrapper[4862]: I1128 11:29:08.853007 4862 scope.go:117] "RemoveContainer" containerID="02d7a90ca0fe317d4dbf751df5688dfd7e5caddce5928ccca5bd882e58b85141" Nov 28 11:29:09 crc kubenswrapper[4862]: I1128 11:29:09.862707 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"6d25f125eb927ad9f5d5ebf41c0bfc64bfcc8ea0c1567eba44d6b656d68eb6c7"} Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.215962 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp"] Nov 28 11:30:00 crc kubenswrapper[4862]: E1128 11:30:00.217233 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e143937-4cf3-4fa0-aa09-381d8bf99f16" containerName="registry" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.217258 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e143937-4cf3-4fa0-aa09-381d8bf99f16" containerName="registry" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.217457 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e143937-4cf3-4fa0-aa09-381d8bf99f16" containerName="registry" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.218336 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.221933 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.221945 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.226433 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp"] Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.348729 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/022a92f9-136d-47c3-891b-71c8b2146c80-config-volume\") pod \"collect-profiles-29405490-gfgpp\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.348823 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/022a92f9-136d-47c3-891b-71c8b2146c80-secret-volume\") pod \"collect-profiles-29405490-gfgpp\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.349015 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25tht\" (UniqueName: \"kubernetes.io/projected/022a92f9-136d-47c3-891b-71c8b2146c80-kube-api-access-25tht\") pod \"collect-profiles-29405490-gfgpp\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.450623 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/022a92f9-136d-47c3-891b-71c8b2146c80-config-volume\") pod \"collect-profiles-29405490-gfgpp\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.450711 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/022a92f9-136d-47c3-891b-71c8b2146c80-secret-volume\") pod \"collect-profiles-29405490-gfgpp\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.450803 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25tht\" (UniqueName: \"kubernetes.io/projected/022a92f9-136d-47c3-891b-71c8b2146c80-kube-api-access-25tht\") pod \"collect-profiles-29405490-gfgpp\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.451805 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/022a92f9-136d-47c3-891b-71c8b2146c80-config-volume\") pod \"collect-profiles-29405490-gfgpp\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.461964 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/022a92f9-136d-47c3-891b-71c8b2146c80-secret-volume\") pod \"collect-profiles-29405490-gfgpp\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.490856 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25tht\" (UniqueName: \"kubernetes.io/projected/022a92f9-136d-47c3-891b-71c8b2146c80-kube-api-access-25tht\") pod \"collect-profiles-29405490-gfgpp\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.547459 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:00 crc kubenswrapper[4862]: I1128 11:30:00.828389 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp"] Nov 28 11:30:00 crc kubenswrapper[4862]: W1128 11:30:00.834359 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod022a92f9_136d_47c3_891b_71c8b2146c80.slice/crio-8fa5ab0b0ff086adaa8e8f63e8375c52cc4940f4d6ee73b66c9078d1dfbee298 WatchSource:0}: Error finding container 8fa5ab0b0ff086adaa8e8f63e8375c52cc4940f4d6ee73b66c9078d1dfbee298: Status 404 returned error can't find the container with id 8fa5ab0b0ff086adaa8e8f63e8375c52cc4940f4d6ee73b66c9078d1dfbee298 Nov 28 11:30:01 crc kubenswrapper[4862]: I1128 11:30:01.237490 4862 generic.go:334] "Generic (PLEG): container finished" podID="022a92f9-136d-47c3-891b-71c8b2146c80" containerID="3ba8f23b2e3be67d75bf070bb36c364c7ec85dfc8ff0a53050d79340065e434d" exitCode=0 Nov 28 11:30:01 crc kubenswrapper[4862]: I1128 11:30:01.237535 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" event={"ID":"022a92f9-136d-47c3-891b-71c8b2146c80","Type":"ContainerDied","Data":"3ba8f23b2e3be67d75bf070bb36c364c7ec85dfc8ff0a53050d79340065e434d"} Nov 28 11:30:01 crc kubenswrapper[4862]: I1128 11:30:01.237573 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" event={"ID":"022a92f9-136d-47c3-891b-71c8b2146c80","Type":"ContainerStarted","Data":"8fa5ab0b0ff086adaa8e8f63e8375c52cc4940f4d6ee73b66c9078d1dfbee298"} Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.549727 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.684873 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/022a92f9-136d-47c3-891b-71c8b2146c80-config-volume\") pod \"022a92f9-136d-47c3-891b-71c8b2146c80\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.684956 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/022a92f9-136d-47c3-891b-71c8b2146c80-secret-volume\") pod \"022a92f9-136d-47c3-891b-71c8b2146c80\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.685000 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25tht\" (UniqueName: \"kubernetes.io/projected/022a92f9-136d-47c3-891b-71c8b2146c80-kube-api-access-25tht\") pod \"022a92f9-136d-47c3-891b-71c8b2146c80\" (UID: \"022a92f9-136d-47c3-891b-71c8b2146c80\") " Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.687331 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/022a92f9-136d-47c3-891b-71c8b2146c80-config-volume" (OuterVolumeSpecName: "config-volume") pod "022a92f9-136d-47c3-891b-71c8b2146c80" (UID: "022a92f9-136d-47c3-891b-71c8b2146c80"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.693490 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/022a92f9-136d-47c3-891b-71c8b2146c80-kube-api-access-25tht" (OuterVolumeSpecName: "kube-api-access-25tht") pod "022a92f9-136d-47c3-891b-71c8b2146c80" (UID: "022a92f9-136d-47c3-891b-71c8b2146c80"). InnerVolumeSpecName "kube-api-access-25tht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.694144 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/022a92f9-136d-47c3-891b-71c8b2146c80-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "022a92f9-136d-47c3-891b-71c8b2146c80" (UID: "022a92f9-136d-47c3-891b-71c8b2146c80"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.786560 4862 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/022a92f9-136d-47c3-891b-71c8b2146c80-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.786595 4862 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/022a92f9-136d-47c3-891b-71c8b2146c80-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 11:30:02 crc kubenswrapper[4862]: I1128 11:30:02.786607 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25tht\" (UniqueName: \"kubernetes.io/projected/022a92f9-136d-47c3-891b-71c8b2146c80-kube-api-access-25tht\") on node \"crc\" DevicePath \"\"" Nov 28 11:30:03 crc kubenswrapper[4862]: I1128 11:30:03.255685 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" event={"ID":"022a92f9-136d-47c3-891b-71c8b2146c80","Type":"ContainerDied","Data":"8fa5ab0b0ff086adaa8e8f63e8375c52cc4940f4d6ee73b66c9078d1dfbee298"} Nov 28 11:30:03 crc kubenswrapper[4862]: I1128 11:30:03.256588 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fa5ab0b0ff086adaa8e8f63e8375c52cc4940f4d6ee73b66c9078d1dfbee298" Nov 28 11:30:03 crc kubenswrapper[4862]: I1128 11:30:03.255799 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405490-gfgpp" Nov 28 11:31:38 crc kubenswrapper[4862]: I1128 11:31:38.291721 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:31:38 crc kubenswrapper[4862]: I1128 11:31:38.292496 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:32:08 crc kubenswrapper[4862]: I1128 11:32:08.292331 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:32:08 crc kubenswrapper[4862]: I1128 11:32:08.293324 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:32:38 crc kubenswrapper[4862]: I1128 11:32:38.292526 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:32:38 crc kubenswrapper[4862]: I1128 11:32:38.293396 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:32:38 crc kubenswrapper[4862]: I1128 11:32:38.293468 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:32:38 crc kubenswrapper[4862]: I1128 11:32:38.294371 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6d25f125eb927ad9f5d5ebf41c0bfc64bfcc8ea0c1567eba44d6b656d68eb6c7"} pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:32:38 crc kubenswrapper[4862]: I1128 11:32:38.294472 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" containerID="cri-o://6d25f125eb927ad9f5d5ebf41c0bfc64bfcc8ea0c1567eba44d6b656d68eb6c7" gracePeriod=600 Nov 28 11:32:39 crc kubenswrapper[4862]: I1128 11:32:39.431742 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6845439-1a4c-4f28-9ba4-29149d545995" containerID="6d25f125eb927ad9f5d5ebf41c0bfc64bfcc8ea0c1567eba44d6b656d68eb6c7" exitCode=0 Nov 28 11:32:39 crc kubenswrapper[4862]: I1128 11:32:39.431802 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerDied","Data":"6d25f125eb927ad9f5d5ebf41c0bfc64bfcc8ea0c1567eba44d6b656d68eb6c7"} Nov 28 11:32:39 crc kubenswrapper[4862]: I1128 11:32:39.432250 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"cd460d7f02e8c5ea1bc5f7ee547d7f0b295f608f496d14e974dc72d5aea4821d"} Nov 28 11:32:39 crc kubenswrapper[4862]: I1128 11:32:39.432289 4862 scope.go:117] "RemoveContainer" containerID="b510eb115f59f27520a8dc8f9d8b64d0f26aef3c7c9a1efef94445ccf2f2eb72" Nov 28 11:34:35 crc kubenswrapper[4862]: I1128 11:34:35.497142 4862 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 11:34:58 crc kubenswrapper[4862]: I1128 11:34:58.825445 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ccpp6"] Nov 28 11:34:58 crc kubenswrapper[4862]: I1128 11:34:58.835073 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovn-controller" containerID="cri-o://43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce" gracePeriod=30 Nov 28 11:34:58 crc kubenswrapper[4862]: I1128 11:34:58.835580 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="northd" containerID="cri-o://bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2" gracePeriod=30 Nov 28 11:34:58 crc kubenswrapper[4862]: I1128 11:34:58.835717 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e" gracePeriod=30 Nov 28 11:34:58 crc kubenswrapper[4862]: I1128 11:34:58.835815 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kube-rbac-proxy-node" containerID="cri-o://7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1" gracePeriod=30 Nov 28 11:34:58 crc kubenswrapper[4862]: I1128 11:34:58.835877 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovn-acl-logging" containerID="cri-o://5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817" gracePeriod=30 Nov 28 11:34:58 crc kubenswrapper[4862]: I1128 11:34:58.835917 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="sbdb" containerID="cri-o://23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5" gracePeriod=30 Nov 28 11:34:58 crc kubenswrapper[4862]: I1128 11:34:58.835990 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="nbdb" containerID="cri-o://33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761" gracePeriod=30 Nov 28 11:34:58 crc kubenswrapper[4862]: I1128 11:34:58.893534 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" containerID="cri-o://d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e" gracePeriod=30 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.192528 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/3.log" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.195530 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovn-acl-logging/0.log" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.196040 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovn-controller/0.log" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.196542 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268240 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lgs2j"] Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268478 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kubecfg-setup" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268495 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kubecfg-setup" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268508 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268518 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268531 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="sbdb" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268539 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="sbdb" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268550 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268558 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268567 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="nbdb" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268574 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="nbdb" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268584 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovn-acl-logging" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268591 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovn-acl-logging" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268602 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268610 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268620 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268627 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268640 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268648 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268657 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="northd" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268664 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="northd" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268677 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kube-rbac-proxy-node" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268685 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kube-rbac-proxy-node" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268697 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovn-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268705 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovn-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268718 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="022a92f9-136d-47c3-891b-71c8b2146c80" containerName="collect-profiles" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268725 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="022a92f9-136d-47c3-891b-71c8b2146c80" containerName="collect-profiles" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.268734 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268742 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268861 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kube-rbac-proxy-node" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268874 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268884 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="northd" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268898 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268909 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="022a92f9-136d-47c3-891b-71c8b2146c80" containerName="collect-profiles" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268918 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268926 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovn-acl-logging" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268934 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268944 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268954 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovnkube-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268964 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="ovn-controller" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268972 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="nbdb" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.268981 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerName="sbdb" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.271392 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293255 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-log-socket\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293312 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9916d783-c2df-47fa-9c38-ee7537db24d4-ovn-node-metrics-cert\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293360 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-var-lib-openvswitch\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293381 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-log-socket" (OuterVolumeSpecName: "log-socket") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293405 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-node-log\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293436 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-openvswitch\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293440 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293487 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-script-lib\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293523 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-kubelet\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293518 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-node-log" (OuterVolumeSpecName: "node-log") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293552 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293572 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-systemd-units\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293602 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293604 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-netns\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293639 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293680 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293687 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293719 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293736 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjwcr\" (UniqueName: \"kubernetes.io/projected/9916d783-c2df-47fa-9c38-ee7537db24d4-kube-api-access-tjwcr\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293779 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-ovn\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293809 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-bin\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293875 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-slash\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293907 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-config\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293942 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-systemd\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.293989 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-env-overrides\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294023 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-ovn-kubernetes\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294067 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-netd\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294121 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-etc-openvswitch\") pod \"9916d783-c2df-47fa-9c38-ee7537db24d4\" (UID: \"9916d783-c2df-47fa-9c38-ee7537db24d4\") " Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294442 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294660 4862 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-log-socket\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294684 4862 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294684 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294703 4862 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-node-log\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294737 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294742 4862 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294775 4862 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294793 4862 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294809 4862 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294827 4862 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294849 4862 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294891 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294938 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.294973 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.295011 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-slash" (OuterVolumeSpecName: "host-slash") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.295324 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.295733 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.299628 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9916d783-c2df-47fa-9c38-ee7537db24d4-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.299798 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9916d783-c2df-47fa-9c38-ee7537db24d4-kube-api-access-tjwcr" (OuterVolumeSpecName: "kube-api-access-tjwcr") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "kube-api-access-tjwcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.311876 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9916d783-c2df-47fa-9c38-ee7537db24d4" (UID: "9916d783-c2df-47fa-9c38-ee7537db24d4"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.395934 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-run-openvswitch\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.395990 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-log-socket\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396029 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-cni-netd\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396061 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-slash\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396134 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/77a3681c-a80f-488f-9c42-4ed8cf095dd1-ovnkube-script-lib\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396178 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-node-log\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396210 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-cni-bin\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396234 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-run-ovn\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396343 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-var-lib-openvswitch\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396404 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/77a3681c-a80f-488f-9c42-4ed8cf095dd1-ovnkube-config\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396428 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/77a3681c-a80f-488f-9c42-4ed8cf095dd1-env-overrides\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396455 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-systemd-units\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396479 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-run-netns\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396538 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d65v\" (UniqueName: \"kubernetes.io/projected/77a3681c-a80f-488f-9c42-4ed8cf095dd1-kube-api-access-2d65v\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396586 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-etc-openvswitch\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396620 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-run-systemd\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396639 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-run-ovn-kubernetes\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396673 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/77a3681c-a80f-488f-9c42-4ed8cf095dd1-ovn-node-metrics-cert\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396702 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396737 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-kubelet\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396818 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjwcr\" (UniqueName: \"kubernetes.io/projected/9916d783-c2df-47fa-9c38-ee7537db24d4-kube-api-access-tjwcr\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396831 4862 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396843 4862 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396854 4862 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-slash\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396866 4862 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396880 4862 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396891 4862 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9916d783-c2df-47fa-9c38-ee7537db24d4-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396902 4862 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396913 4862 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396924 4862 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9916d783-c2df-47fa-9c38-ee7537db24d4-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.396935 4862 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9916d783-c2df-47fa-9c38-ee7537db24d4-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.446504 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kqpd4_d6abffd5-6086-4a32-ba8d-e7c669548c5f/kube-multus/2.log" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.447317 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kqpd4_d6abffd5-6086-4a32-ba8d-e7c669548c5f/kube-multus/1.log" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.447397 4862 generic.go:334] "Generic (PLEG): container finished" podID="d6abffd5-6086-4a32-ba8d-e7c669548c5f" containerID="e805b51f4db68a36820f093a795585bd38dc9ad4bb444695e78ebef9a461111e" exitCode=2 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.447498 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kqpd4" event={"ID":"d6abffd5-6086-4a32-ba8d-e7c669548c5f","Type":"ContainerDied","Data":"e805b51f4db68a36820f093a795585bd38dc9ad4bb444695e78ebef9a461111e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.447554 4862 scope.go:117] "RemoveContainer" containerID="743ff03be2e17e9299a44689ca38866a4f301f27a42048cce4252a8cdc7cc625" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.448421 4862 scope.go:117] "RemoveContainer" containerID="e805b51f4db68a36820f093a795585bd38dc9ad4bb444695e78ebef9a461111e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.452068 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovnkube-controller/3.log" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.461581 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovn-acl-logging/0.log" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.463473 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ccpp6_9916d783-c2df-47fa-9c38-ee7537db24d4/ovn-controller/0.log" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464084 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e" exitCode=0 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464175 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5" exitCode=0 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464200 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761" exitCode=0 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464223 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2" exitCode=0 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464242 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e" exitCode=0 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464260 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1" exitCode=0 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464278 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817" exitCode=143 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464298 4862 generic.go:334] "Generic (PLEG): container finished" podID="9916d783-c2df-47fa-9c38-ee7537db24d4" containerID="43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce" exitCode=143 Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464338 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464394 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464430 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464451 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464460 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464773 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464802 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464818 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464833 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464843 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464850 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464858 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464865 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464872 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464879 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464886 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464893 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464903 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464913 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464922 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464929 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464937 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464945 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464952 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464959 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464965 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464973 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464981 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.464990 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465000 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465009 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465015 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465022 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465028 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465036 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465042 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465049 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465055 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465061 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465070 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ccpp6" event={"ID":"9916d783-c2df-47fa-9c38-ee7537db24d4","Type":"ContainerDied","Data":"7a759715337151e5488c580bd6e803f7ea34cd6c692e8c9775e9ab62a62bdce5"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465081 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465117 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465138 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465145 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465151 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465158 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465165 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465171 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465179 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.465186 4862 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69"} Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.493032 4862 scope.go:117] "RemoveContainer" containerID="d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.497910 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-node-log\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.497951 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-cni-bin\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.497977 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-run-ovn\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498000 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-var-lib-openvswitch\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498024 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/77a3681c-a80f-488f-9c42-4ed8cf095dd1-ovnkube-config\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498049 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/77a3681c-a80f-488f-9c42-4ed8cf095dd1-env-overrides\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498074 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-systemd-units\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498153 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-run-netns\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498229 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d65v\" (UniqueName: \"kubernetes.io/projected/77a3681c-a80f-488f-9c42-4ed8cf095dd1-kube-api-access-2d65v\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498267 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-var-lib-openvswitch\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498320 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-etc-openvswitch\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498364 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-node-log\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498406 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-cni-bin\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498444 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-run-ovn\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498276 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-etc-openvswitch\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498493 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-run-systemd\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498524 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-run-ovn-kubernetes\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498558 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/77a3681c-a80f-488f-9c42-4ed8cf095dd1-ovn-node-metrics-cert\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498592 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498608 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-systemd-units\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498630 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-kubelet\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498685 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-run-openvswitch\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498714 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-log-socket\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498746 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-cni-netd\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498787 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-slash\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498821 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/77a3681c-a80f-488f-9c42-4ed8cf095dd1-ovnkube-script-lib\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498858 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498940 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-cni-netd\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498939 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-run-ovn-kubernetes\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498934 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-log-socket\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498961 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-slash\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498984 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-kubelet\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.499015 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-run-openvswitch\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.498997 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-run-systemd\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.499223 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/77a3681c-a80f-488f-9c42-4ed8cf095dd1-env-overrides\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.499279 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/77a3681c-a80f-488f-9c42-4ed8cf095dd1-host-run-netns\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.500001 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/77a3681c-a80f-488f-9c42-4ed8cf095dd1-ovnkube-script-lib\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.501287 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/77a3681c-a80f-488f-9c42-4ed8cf095dd1-ovnkube-config\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.505068 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/77a3681c-a80f-488f-9c42-4ed8cf095dd1-ovn-node-metrics-cert\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.529138 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d65v\" (UniqueName: \"kubernetes.io/projected/77a3681c-a80f-488f-9c42-4ed8cf095dd1-kube-api-access-2d65v\") pod \"ovnkube-node-lgs2j\" (UID: \"77a3681c-a80f-488f-9c42-4ed8cf095dd1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.546384 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.548018 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ccpp6"] Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.552068 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ccpp6"] Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.582598 4862 scope.go:117] "RemoveContainer" containerID="23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.593374 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.604331 4862 scope.go:117] "RemoveContainer" containerID="33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.635739 4862 scope.go:117] "RemoveContainer" containerID="bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2" Nov 28 11:34:59 crc kubenswrapper[4862]: W1128 11:34:59.647512 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77a3681c_a80f_488f_9c42_4ed8cf095dd1.slice/crio-7d21185f0bb1a979214482d7fdafe709b0f362fd4d6bc5af1f2ebc33e95a91be WatchSource:0}: Error finding container 7d21185f0bb1a979214482d7fdafe709b0f362fd4d6bc5af1f2ebc33e95a91be: Status 404 returned error can't find the container with id 7d21185f0bb1a979214482d7fdafe709b0f362fd4d6bc5af1f2ebc33e95a91be Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.655189 4862 scope.go:117] "RemoveContainer" containerID="9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.676992 4862 scope.go:117] "RemoveContainer" containerID="7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.694636 4862 scope.go:117] "RemoveContainer" containerID="5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.721771 4862 scope.go:117] "RemoveContainer" containerID="43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.745173 4862 scope.go:117] "RemoveContainer" containerID="2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.768884 4862 scope.go:117] "RemoveContainer" containerID="d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.770597 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e\": container with ID starting with d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e not found: ID does not exist" containerID="d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.770637 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} err="failed to get container status \"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e\": rpc error: code = NotFound desc = could not find container \"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e\": container with ID starting with d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.770672 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.772038 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\": container with ID starting with 80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef not found: ID does not exist" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.772148 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} err="failed to get container status \"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\": rpc error: code = NotFound desc = could not find container \"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\": container with ID starting with 80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.772190 4862 scope.go:117] "RemoveContainer" containerID="23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.781082 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\": container with ID starting with 23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5 not found: ID does not exist" containerID="23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.781159 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} err="failed to get container status \"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\": rpc error: code = NotFound desc = could not find container \"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\": container with ID starting with 23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.781199 4862 scope.go:117] "RemoveContainer" containerID="33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.782009 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\": container with ID starting with 33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761 not found: ID does not exist" containerID="33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.782046 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} err="failed to get container status \"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\": rpc error: code = NotFound desc = could not find container \"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\": container with ID starting with 33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.782063 4862 scope.go:117] "RemoveContainer" containerID="bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.782669 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\": container with ID starting with bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2 not found: ID does not exist" containerID="bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.782689 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} err="failed to get container status \"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\": rpc error: code = NotFound desc = could not find container \"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\": container with ID starting with bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.782705 4862 scope.go:117] "RemoveContainer" containerID="9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.791134 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\": container with ID starting with 9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e not found: ID does not exist" containerID="9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.791188 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} err="failed to get container status \"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\": rpc error: code = NotFound desc = could not find container \"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\": container with ID starting with 9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.791229 4862 scope.go:117] "RemoveContainer" containerID="7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.791906 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\": container with ID starting with 7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1 not found: ID does not exist" containerID="7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.791954 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} err="failed to get container status \"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\": rpc error: code = NotFound desc = could not find container \"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\": container with ID starting with 7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.791991 4862 scope.go:117] "RemoveContainer" containerID="5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.792297 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\": container with ID starting with 5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817 not found: ID does not exist" containerID="5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.792330 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} err="failed to get container status \"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\": rpc error: code = NotFound desc = could not find container \"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\": container with ID starting with 5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.792353 4862 scope.go:117] "RemoveContainer" containerID="43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.793604 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\": container with ID starting with 43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce not found: ID does not exist" containerID="43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.793649 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} err="failed to get container status \"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\": rpc error: code = NotFound desc = could not find container \"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\": container with ID starting with 43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.793674 4862 scope.go:117] "RemoveContainer" containerID="2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69" Nov 28 11:34:59 crc kubenswrapper[4862]: E1128 11:34:59.794558 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\": container with ID starting with 2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69 not found: ID does not exist" containerID="2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.794596 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69"} err="failed to get container status \"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\": rpc error: code = NotFound desc = could not find container \"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\": container with ID starting with 2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.794638 4862 scope.go:117] "RemoveContainer" containerID="d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.795033 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} err="failed to get container status \"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e\": rpc error: code = NotFound desc = could not find container \"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e\": container with ID starting with d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.795065 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.796548 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} err="failed to get container status \"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\": rpc error: code = NotFound desc = could not find container \"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\": container with ID starting with 80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.796567 4862 scope.go:117] "RemoveContainer" containerID="23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.796883 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} err="failed to get container status \"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\": rpc error: code = NotFound desc = could not find container \"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\": container with ID starting with 23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.796913 4862 scope.go:117] "RemoveContainer" containerID="33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.797243 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} err="failed to get container status \"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\": rpc error: code = NotFound desc = could not find container \"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\": container with ID starting with 33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.797278 4862 scope.go:117] "RemoveContainer" containerID="bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.797639 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} err="failed to get container status \"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\": rpc error: code = NotFound desc = could not find container \"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\": container with ID starting with bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.797656 4862 scope.go:117] "RemoveContainer" containerID="9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.797888 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} err="failed to get container status \"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\": rpc error: code = NotFound desc = could not find container \"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\": container with ID starting with 9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.797905 4862 scope.go:117] "RemoveContainer" containerID="7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.798336 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} err="failed to get container status \"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\": rpc error: code = NotFound desc = could not find container \"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\": container with ID starting with 7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.798355 4862 scope.go:117] "RemoveContainer" containerID="5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.798560 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} err="failed to get container status \"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\": rpc error: code = NotFound desc = could not find container \"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\": container with ID starting with 5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.798581 4862 scope.go:117] "RemoveContainer" containerID="43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.798752 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} err="failed to get container status \"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\": rpc error: code = NotFound desc = could not find container \"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\": container with ID starting with 43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.798767 4862 scope.go:117] "RemoveContainer" containerID="2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.798935 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69"} err="failed to get container status \"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\": rpc error: code = NotFound desc = could not find container \"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\": container with ID starting with 2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.798952 4862 scope.go:117] "RemoveContainer" containerID="d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.799150 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} err="failed to get container status \"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e\": rpc error: code = NotFound desc = could not find container \"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e\": container with ID starting with d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.799167 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.799450 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} err="failed to get container status \"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\": rpc error: code = NotFound desc = could not find container \"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\": container with ID starting with 80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.799481 4862 scope.go:117] "RemoveContainer" containerID="23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.799817 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} err="failed to get container status \"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\": rpc error: code = NotFound desc = could not find container \"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\": container with ID starting with 23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.799855 4862 scope.go:117] "RemoveContainer" containerID="33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.800167 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} err="failed to get container status \"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\": rpc error: code = NotFound desc = could not find container \"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\": container with ID starting with 33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.800228 4862 scope.go:117] "RemoveContainer" containerID="bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.800484 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} err="failed to get container status \"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\": rpc error: code = NotFound desc = could not find container \"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\": container with ID starting with bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.800506 4862 scope.go:117] "RemoveContainer" containerID="9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.800795 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} err="failed to get container status \"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\": rpc error: code = NotFound desc = could not find container \"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\": container with ID starting with 9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.800822 4862 scope.go:117] "RemoveContainer" containerID="7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.801056 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} err="failed to get container status \"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\": rpc error: code = NotFound desc = could not find container \"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\": container with ID starting with 7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.801085 4862 scope.go:117] "RemoveContainer" containerID="5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.801342 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} err="failed to get container status \"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\": rpc error: code = NotFound desc = could not find container \"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\": container with ID starting with 5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.801374 4862 scope.go:117] "RemoveContainer" containerID="43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.801617 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} err="failed to get container status \"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\": rpc error: code = NotFound desc = could not find container \"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\": container with ID starting with 43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.801654 4862 scope.go:117] "RemoveContainer" containerID="2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.801931 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69"} err="failed to get container status \"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\": rpc error: code = NotFound desc = could not find container \"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\": container with ID starting with 2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.801997 4862 scope.go:117] "RemoveContainer" containerID="d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.802322 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e"} err="failed to get container status \"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e\": rpc error: code = NotFound desc = could not find container \"d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e\": container with ID starting with d6ac7e3fd35c1778bd5fa15af9e7e8ef231c44104aa3829b0c6abb10fa74510e not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.802354 4862 scope.go:117] "RemoveContainer" containerID="80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.802562 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef"} err="failed to get container status \"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\": rpc error: code = NotFound desc = could not find container \"80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef\": container with ID starting with 80d2b8759f9f93d544b5545c3c2bb665e0bd99573d3e5b05a396101cc33aceef not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.802580 4862 scope.go:117] "RemoveContainer" containerID="23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.802754 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5"} err="failed to get container status \"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\": rpc error: code = NotFound desc = could not find container \"23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5\": container with ID starting with 23bfadc4c1d53a329d7143b5b9a0f722607928f32e32b5bb1c47292ed0b6d1c5 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.802779 4862 scope.go:117] "RemoveContainer" containerID="33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.803317 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761"} err="failed to get container status \"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\": rpc error: code = NotFound desc = could not find container \"33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761\": container with ID starting with 33db99e20ad00b68992413b1abb0d3cdd70400bb040c084a786afa987dd70761 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.803372 4862 scope.go:117] "RemoveContainer" containerID="bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.803812 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2"} err="failed to get container status \"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\": rpc error: code = NotFound desc = could not find container \"bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2\": container with ID starting with bbd5e2f2993c5b2a424c5fc00462d6382d8a725cabd1bda262ae5f3df25b1fc2 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.803829 4862 scope.go:117] "RemoveContainer" containerID="9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.808370 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e"} err="failed to get container status \"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\": rpc error: code = NotFound desc = could not find container \"9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e\": container with ID starting with 9ab2ebf38153001b032f781a3b9bd9b03b8b962207905a37d7a6e19cb4ea1b1e not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.808401 4862 scope.go:117] "RemoveContainer" containerID="7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.808723 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1"} err="failed to get container status \"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\": rpc error: code = NotFound desc = could not find container \"7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1\": container with ID starting with 7a0eb27a2871596c4af968c2790d31750d08604282bf9928801e05c43f23adb1 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.808772 4862 scope.go:117] "RemoveContainer" containerID="5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.809230 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817"} err="failed to get container status \"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\": rpc error: code = NotFound desc = could not find container \"5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817\": container with ID starting with 5918f538270a4d0464be49732faba5a54bc94877d0419f699d3163fe0e92d817 not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.809256 4862 scope.go:117] "RemoveContainer" containerID="43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.809607 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce"} err="failed to get container status \"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\": rpc error: code = NotFound desc = could not find container \"43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce\": container with ID starting with 43b500a59b49a3a8a0442bd0aafdc4c6915704e470cd20fc62721b09d70f1bce not found: ID does not exist" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.809624 4862 scope.go:117] "RemoveContainer" containerID="2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69" Nov 28 11:34:59 crc kubenswrapper[4862]: I1128 11:34:59.810162 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69"} err="failed to get container status \"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\": rpc error: code = NotFound desc = could not find container \"2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69\": container with ID starting with 2a270a6efc8b869a423d235a69b9a9e045f48c6714200f60c3664a7a1f4abc69 not found: ID does not exist" Nov 28 11:35:00 crc kubenswrapper[4862]: I1128 11:35:00.474883 4862 generic.go:334] "Generic (PLEG): container finished" podID="77a3681c-a80f-488f-9c42-4ed8cf095dd1" containerID="b119cb016eebd8d47a5dec6ed9b2b5ef5f0755e4c7fee9fc60576ad50e5058df" exitCode=0 Nov 28 11:35:00 crc kubenswrapper[4862]: I1128 11:35:00.475017 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerDied","Data":"b119cb016eebd8d47a5dec6ed9b2b5ef5f0755e4c7fee9fc60576ad50e5058df"} Nov 28 11:35:00 crc kubenswrapper[4862]: I1128 11:35:00.475375 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerStarted","Data":"7d21185f0bb1a979214482d7fdafe709b0f362fd4d6bc5af1f2ebc33e95a91be"} Nov 28 11:35:00 crc kubenswrapper[4862]: I1128 11:35:00.480578 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kqpd4_d6abffd5-6086-4a32-ba8d-e7c669548c5f/kube-multus/2.log" Nov 28 11:35:00 crc kubenswrapper[4862]: I1128 11:35:00.480677 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kqpd4" event={"ID":"d6abffd5-6086-4a32-ba8d-e7c669548c5f","Type":"ContainerStarted","Data":"d34bd72945cca4b299b098949e133e2e2ea819efe64bd05778d6dee247b41742"} Nov 28 11:35:00 crc kubenswrapper[4862]: I1128 11:35:00.847010 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9916d783-c2df-47fa-9c38-ee7537db24d4" path="/var/lib/kubelet/pods/9916d783-c2df-47fa-9c38-ee7537db24d4/volumes" Nov 28 11:35:01 crc kubenswrapper[4862]: I1128 11:35:01.490984 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerStarted","Data":"1006e22b2eba53f10b3bb5a92e98df18192684895bc32135f652fe6195d9034d"} Nov 28 11:35:01 crc kubenswrapper[4862]: I1128 11:35:01.491629 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerStarted","Data":"a33db927337c7824a8020797efe7426bb94839e62e1255c8e444eb715ac78030"} Nov 28 11:35:01 crc kubenswrapper[4862]: I1128 11:35:01.491651 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerStarted","Data":"8b905e1a0bf16298fd7c667ee4acf8d675ba28adf827fa7a7e13527da9a1e599"} Nov 28 11:35:01 crc kubenswrapper[4862]: I1128 11:35:01.491672 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerStarted","Data":"628c61a014b21e8bcc01ab354deb9d9e34e4d93a230f9956ed0c88483e84b85a"} Nov 28 11:35:01 crc kubenswrapper[4862]: I1128 11:35:01.491689 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerStarted","Data":"fed78302100b28ab68da5e4682c48813e0ddeab130bae0c4e15e097f1111d646"} Nov 28 11:35:01 crc kubenswrapper[4862]: I1128 11:35:01.491705 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerStarted","Data":"db99f50c5febf3a60fe614ab7ff5e005fcda3079a5d93fe8c53d483fc470b0de"} Nov 28 11:35:04 crc kubenswrapper[4862]: I1128 11:35:04.535282 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerStarted","Data":"ec97a561589b34d4197d26c68b4923000769f1a35a685b8d11588864a7ecf3f3"} Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.402811 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-jxc9f"] Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.408626 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.412587 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.412613 4862 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-7hxgs" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.412667 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.412599 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.418188 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-node-mnt\") pod \"crc-storage-crc-jxc9f\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.418358 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-crc-storage\") pod \"crc-storage-crc-jxc9f\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.418454 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb7bx\" (UniqueName: \"kubernetes.io/projected/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-kube-api-access-lb7bx\") pod \"crc-storage-crc-jxc9f\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.519761 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-node-mnt\") pod \"crc-storage-crc-jxc9f\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.519855 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-crc-storage\") pod \"crc-storage-crc-jxc9f\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.519896 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb7bx\" (UniqueName: \"kubernetes.io/projected/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-kube-api-access-lb7bx\") pod \"crc-storage-crc-jxc9f\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.521151 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-crc-storage\") pod \"crc-storage-crc-jxc9f\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.521283 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-node-mnt\") pod \"crc-storage-crc-jxc9f\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.557994 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb7bx\" (UniqueName: \"kubernetes.io/projected/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-kube-api-access-lb7bx\") pod \"crc-storage-crc-jxc9f\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.568175 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" event={"ID":"77a3681c-a80f-488f-9c42-4ed8cf095dd1","Type":"ContainerStarted","Data":"34c6487c165d20d9a084a2b4c7ecd1af9a31e994c074dc4cd4fc1d5fc2a33ae3"} Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.661793 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-jxc9f"] Nov 28 11:35:07 crc kubenswrapper[4862]: I1128 11:35:07.724122 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: E1128 11:35:07.749277 4862 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-jxc9f_crc-storage_983ca19c-bbc7-4b2c-882e-1d35a9f43cb9_0(d94a433a8e6bc93a04397523816987482a241462cf05d07cd422291d925e400d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 11:35:07 crc kubenswrapper[4862]: E1128 11:35:07.749357 4862 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-jxc9f_crc-storage_983ca19c-bbc7-4b2c-882e-1d35a9f43cb9_0(d94a433a8e6bc93a04397523816987482a241462cf05d07cd422291d925e400d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: E1128 11:35:07.749380 4862 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-jxc9f_crc-storage_983ca19c-bbc7-4b2c-882e-1d35a9f43cb9_0(d94a433a8e6bc93a04397523816987482a241462cf05d07cd422291d925e400d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:07 crc kubenswrapper[4862]: E1128 11:35:07.749437 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-jxc9f_crc-storage(983ca19c-bbc7-4b2c-882e-1d35a9f43cb9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-jxc9f_crc-storage(983ca19c-bbc7-4b2c-882e-1d35a9f43cb9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-jxc9f_crc-storage_983ca19c-bbc7-4b2c-882e-1d35a9f43cb9_0(d94a433a8e6bc93a04397523816987482a241462cf05d07cd422291d925e400d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-jxc9f" podUID="983ca19c-bbc7-4b2c-882e-1d35a9f43cb9" Nov 28 11:35:08 crc kubenswrapper[4862]: I1128 11:35:08.292227 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:35:08 crc kubenswrapper[4862]: I1128 11:35:08.292323 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:35:08 crc kubenswrapper[4862]: I1128 11:35:08.575264 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:08 crc kubenswrapper[4862]: I1128 11:35:08.575989 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:08 crc kubenswrapper[4862]: E1128 11:35:08.622325 4862 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-jxc9f_crc-storage_983ca19c-bbc7-4b2c-882e-1d35a9f43cb9_0(157020421cdc9879598533b9f2dd0be1f1200020077a3e1b8636922fb871fe3c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 11:35:08 crc kubenswrapper[4862]: E1128 11:35:08.622750 4862 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-jxc9f_crc-storage_983ca19c-bbc7-4b2c-882e-1d35a9f43cb9_0(157020421cdc9879598533b9f2dd0be1f1200020077a3e1b8636922fb871fe3c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:08 crc kubenswrapper[4862]: E1128 11:35:08.622795 4862 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-jxc9f_crc-storage_983ca19c-bbc7-4b2c-882e-1d35a9f43cb9_0(157020421cdc9879598533b9f2dd0be1f1200020077a3e1b8636922fb871fe3c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:08 crc kubenswrapper[4862]: E1128 11:35:08.622871 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-jxc9f_crc-storage(983ca19c-bbc7-4b2c-882e-1d35a9f43cb9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-jxc9f_crc-storage(983ca19c-bbc7-4b2c-882e-1d35a9f43cb9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-jxc9f_crc-storage_983ca19c-bbc7-4b2c-882e-1d35a9f43cb9_0(157020421cdc9879598533b9f2dd0be1f1200020077a3e1b8636922fb871fe3c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-jxc9f" podUID="983ca19c-bbc7-4b2c-882e-1d35a9f43cb9" Nov 28 11:35:08 crc kubenswrapper[4862]: I1128 11:35:08.632774 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" podStartSLOduration=9.632738318 podStartE2EDuration="9.632738318s" podCreationTimestamp="2025-11-28 11:34:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:35:08.62713894 +0000 UTC m=+786.159652891" watchObservedRunningTime="2025-11-28 11:35:08.632738318 +0000 UTC m=+786.165252289" Nov 28 11:35:09 crc kubenswrapper[4862]: I1128 11:35:09.594630 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:35:09 crc kubenswrapper[4862]: I1128 11:35:09.594712 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:35:09 crc kubenswrapper[4862]: I1128 11:35:09.595549 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:35:09 crc kubenswrapper[4862]: I1128 11:35:09.644757 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:35:09 crc kubenswrapper[4862]: I1128 11:35:09.647607 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:35:11 crc kubenswrapper[4862]: I1128 11:35:11.629972 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lgs2j" Nov 28 11:35:22 crc kubenswrapper[4862]: I1128 11:35:22.838211 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:22 crc kubenswrapper[4862]: I1128 11:35:22.841588 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:23 crc kubenswrapper[4862]: I1128 11:35:23.087622 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-jxc9f"] Nov 28 11:35:23 crc kubenswrapper[4862]: I1128 11:35:23.095843 4862 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 11:35:23 crc kubenswrapper[4862]: I1128 11:35:23.715717 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-jxc9f" event={"ID":"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9","Type":"ContainerStarted","Data":"d90c4a7d81ef899a0d1afa59e1361aec3b942ab6cd4633700737fc5a9a314093"} Nov 28 11:35:25 crc kubenswrapper[4862]: I1128 11:35:25.733288 4862 generic.go:334] "Generic (PLEG): container finished" podID="983ca19c-bbc7-4b2c-882e-1d35a9f43cb9" containerID="daf993b9f3bb4d73b9687c7415fe5d58029ce60a0131f2f21f50281b985774e8" exitCode=0 Nov 28 11:35:25 crc kubenswrapper[4862]: I1128 11:35:25.733402 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-jxc9f" event={"ID":"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9","Type":"ContainerDied","Data":"daf993b9f3bb4d73b9687c7415fe5d58029ce60a0131f2f21f50281b985774e8"} Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.088210 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.129059 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb7bx\" (UniqueName: \"kubernetes.io/projected/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-kube-api-access-lb7bx\") pod \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.129140 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-crc-storage\") pod \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.129258 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-node-mnt\") pod \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\" (UID: \"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9\") " Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.129423 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "983ca19c-bbc7-4b2c-882e-1d35a9f43cb9" (UID: "983ca19c-bbc7-4b2c-882e-1d35a9f43cb9"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.134299 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-kube-api-access-lb7bx" (OuterVolumeSpecName: "kube-api-access-lb7bx") pod "983ca19c-bbc7-4b2c-882e-1d35a9f43cb9" (UID: "983ca19c-bbc7-4b2c-882e-1d35a9f43cb9"). InnerVolumeSpecName "kube-api-access-lb7bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.149458 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "983ca19c-bbc7-4b2c-882e-1d35a9f43cb9" (UID: "983ca19c-bbc7-4b2c-882e-1d35a9f43cb9"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.231294 4862 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.231347 4862 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.231367 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb7bx\" (UniqueName: \"kubernetes.io/projected/983ca19c-bbc7-4b2c-882e-1d35a9f43cb9-kube-api-access-lb7bx\") on node \"crc\" DevicePath \"\"" Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.751190 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-jxc9f" event={"ID":"983ca19c-bbc7-4b2c-882e-1d35a9f43cb9","Type":"ContainerDied","Data":"d90c4a7d81ef899a0d1afa59e1361aec3b942ab6cd4633700737fc5a9a314093"} Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.751246 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d90c4a7d81ef899a0d1afa59e1361aec3b942ab6cd4633700737fc5a9a314093" Nov 28 11:35:27 crc kubenswrapper[4862]: I1128 11:35:27.751585 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jxc9f" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.323435 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4"] Nov 28 11:35:35 crc kubenswrapper[4862]: E1128 11:35:35.324614 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983ca19c-bbc7-4b2c-882e-1d35a9f43cb9" containerName="storage" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.324635 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="983ca19c-bbc7-4b2c-882e-1d35a9f43cb9" containerName="storage" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.324754 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="983ca19c-bbc7-4b2c-882e-1d35a9f43cb9" containerName="storage" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.325674 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.330462 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.340279 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4"] Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.466742 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.466852 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.466937 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms5cd\" (UniqueName: \"kubernetes.io/projected/e7347f1d-97ee-4391-9d84-08348032071c-kube-api-access-ms5cd\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.568625 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.568720 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.568868 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms5cd\" (UniqueName: \"kubernetes.io/projected/e7347f1d-97ee-4391-9d84-08348032071c-kube-api-access-ms5cd\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.569656 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.569809 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.595754 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms5cd\" (UniqueName: \"kubernetes.io/projected/e7347f1d-97ee-4391-9d84-08348032071c-kube-api-access-ms5cd\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.696483 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:35 crc kubenswrapper[4862]: I1128 11:35:35.924813 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4"] Nov 28 11:35:36 crc kubenswrapper[4862]: I1128 11:35:36.815198 4862 generic.go:334] "Generic (PLEG): container finished" podID="e7347f1d-97ee-4391-9d84-08348032071c" containerID="ed04ea7244a4ac78dd18fb5a0aabd90503092684fb436ee95fe68f514c57de84" exitCode=0 Nov 28 11:35:36 crc kubenswrapper[4862]: I1128 11:35:36.815291 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" event={"ID":"e7347f1d-97ee-4391-9d84-08348032071c","Type":"ContainerDied","Data":"ed04ea7244a4ac78dd18fb5a0aabd90503092684fb436ee95fe68f514c57de84"} Nov 28 11:35:36 crc kubenswrapper[4862]: I1128 11:35:36.815704 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" event={"ID":"e7347f1d-97ee-4391-9d84-08348032071c","Type":"ContainerStarted","Data":"4275f5449787fdcc0508b98983e27bfe0c8bbfd72dfa8a451a03d32ff9cfc87e"} Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.682905 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q8zgx"] Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.684968 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.698145 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8zgx"] Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.802208 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-catalog-content\") pod \"redhat-operators-q8zgx\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.802406 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5txv\" (UniqueName: \"kubernetes.io/projected/bfa1d663-78ab-4131-b299-b93f4108245f-kube-api-access-x5txv\") pod \"redhat-operators-q8zgx\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.802480 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-utilities\") pod \"redhat-operators-q8zgx\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.904016 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-catalog-content\") pod \"redhat-operators-q8zgx\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.904117 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5txv\" (UniqueName: \"kubernetes.io/projected/bfa1d663-78ab-4131-b299-b93f4108245f-kube-api-access-x5txv\") pod \"redhat-operators-q8zgx\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.904157 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-utilities\") pod \"redhat-operators-q8zgx\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.904595 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-catalog-content\") pod \"redhat-operators-q8zgx\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.906244 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-utilities\") pod \"redhat-operators-q8zgx\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:37 crc kubenswrapper[4862]: I1128 11:35:37.923332 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5txv\" (UniqueName: \"kubernetes.io/projected/bfa1d663-78ab-4131-b299-b93f4108245f-kube-api-access-x5txv\") pod \"redhat-operators-q8zgx\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:38 crc kubenswrapper[4862]: I1128 11:35:38.037935 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:38 crc kubenswrapper[4862]: I1128 11:35:38.258373 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8zgx"] Nov 28 11:35:38 crc kubenswrapper[4862]: W1128 11:35:38.263057 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfa1d663_78ab_4131_b299_b93f4108245f.slice/crio-14a4a5fe7d7e237ec1bd446d04bae1af8494f0d327bd59b3afd5bb9be066f0dc WatchSource:0}: Error finding container 14a4a5fe7d7e237ec1bd446d04bae1af8494f0d327bd59b3afd5bb9be066f0dc: Status 404 returned error can't find the container with id 14a4a5fe7d7e237ec1bd446d04bae1af8494f0d327bd59b3afd5bb9be066f0dc Nov 28 11:35:38 crc kubenswrapper[4862]: I1128 11:35:38.292330 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:35:38 crc kubenswrapper[4862]: I1128 11:35:38.292784 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:35:38 crc kubenswrapper[4862]: I1128 11:35:38.827437 4862 generic.go:334] "Generic (PLEG): container finished" podID="e7347f1d-97ee-4391-9d84-08348032071c" containerID="9873fd3bb109467d83e3baf90f06794890afc859229efd88c7745cf2fbfefe1c" exitCode=0 Nov 28 11:35:38 crc kubenswrapper[4862]: I1128 11:35:38.827550 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" event={"ID":"e7347f1d-97ee-4391-9d84-08348032071c","Type":"ContainerDied","Data":"9873fd3bb109467d83e3baf90f06794890afc859229efd88c7745cf2fbfefe1c"} Nov 28 11:35:38 crc kubenswrapper[4862]: I1128 11:35:38.831438 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zgx" event={"ID":"bfa1d663-78ab-4131-b299-b93f4108245f","Type":"ContainerDied","Data":"aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab"} Nov 28 11:35:38 crc kubenswrapper[4862]: I1128 11:35:38.831310 4862 generic.go:334] "Generic (PLEG): container finished" podID="bfa1d663-78ab-4131-b299-b93f4108245f" containerID="aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab" exitCode=0 Nov 28 11:35:38 crc kubenswrapper[4862]: I1128 11:35:38.833415 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zgx" event={"ID":"bfa1d663-78ab-4131-b299-b93f4108245f","Type":"ContainerStarted","Data":"14a4a5fe7d7e237ec1bd446d04bae1af8494f0d327bd59b3afd5bb9be066f0dc"} Nov 28 11:35:39 crc kubenswrapper[4862]: I1128 11:35:39.847015 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zgx" event={"ID":"bfa1d663-78ab-4131-b299-b93f4108245f","Type":"ContainerStarted","Data":"ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e"} Nov 28 11:35:39 crc kubenswrapper[4862]: I1128 11:35:39.852400 4862 generic.go:334] "Generic (PLEG): container finished" podID="e7347f1d-97ee-4391-9d84-08348032071c" containerID="ecc91d924eecf3bd963b52d9eb37b31e268c229befeed38a7a92bb11fd1d2a8a" exitCode=0 Nov 28 11:35:39 crc kubenswrapper[4862]: I1128 11:35:39.852485 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" event={"ID":"e7347f1d-97ee-4391-9d84-08348032071c","Type":"ContainerDied","Data":"ecc91d924eecf3bd963b52d9eb37b31e268c229befeed38a7a92bb11fd1d2a8a"} Nov 28 11:35:40 crc kubenswrapper[4862]: I1128 11:35:40.862789 4862 generic.go:334] "Generic (PLEG): container finished" podID="bfa1d663-78ab-4131-b299-b93f4108245f" containerID="ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e" exitCode=0 Nov 28 11:35:40 crc kubenswrapper[4862]: I1128 11:35:40.862880 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zgx" event={"ID":"bfa1d663-78ab-4131-b299-b93f4108245f","Type":"ContainerDied","Data":"ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e"} Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.162902 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.249957 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-util\") pod \"e7347f1d-97ee-4391-9d84-08348032071c\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.250029 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-bundle\") pod \"e7347f1d-97ee-4391-9d84-08348032071c\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.250207 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms5cd\" (UniqueName: \"kubernetes.io/projected/e7347f1d-97ee-4391-9d84-08348032071c-kube-api-access-ms5cd\") pod \"e7347f1d-97ee-4391-9d84-08348032071c\" (UID: \"e7347f1d-97ee-4391-9d84-08348032071c\") " Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.251079 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-bundle" (OuterVolumeSpecName: "bundle") pod "e7347f1d-97ee-4391-9d84-08348032071c" (UID: "e7347f1d-97ee-4391-9d84-08348032071c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.259860 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7347f1d-97ee-4391-9d84-08348032071c-kube-api-access-ms5cd" (OuterVolumeSpecName: "kube-api-access-ms5cd") pod "e7347f1d-97ee-4391-9d84-08348032071c" (UID: "e7347f1d-97ee-4391-9d84-08348032071c"). InnerVolumeSpecName "kube-api-access-ms5cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.264601 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-util" (OuterVolumeSpecName: "util") pod "e7347f1d-97ee-4391-9d84-08348032071c" (UID: "e7347f1d-97ee-4391-9d84-08348032071c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.352600 4862 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.352708 4862 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7347f1d-97ee-4391-9d84-08348032071c-util\") on node \"crc\" DevicePath \"\"" Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.352741 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms5cd\" (UniqueName: \"kubernetes.io/projected/e7347f1d-97ee-4391-9d84-08348032071c-kube-api-access-ms5cd\") on node \"crc\" DevicePath \"\"" Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.878552 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zgx" event={"ID":"bfa1d663-78ab-4131-b299-b93f4108245f","Type":"ContainerStarted","Data":"7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c"} Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.885530 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" event={"ID":"e7347f1d-97ee-4391-9d84-08348032071c","Type":"ContainerDied","Data":"4275f5449787fdcc0508b98983e27bfe0c8bbfd72dfa8a451a03d32ff9cfc87e"} Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.885589 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4275f5449787fdcc0508b98983e27bfe0c8bbfd72dfa8a451a03d32ff9cfc87e" Nov 28 11:35:41 crc kubenswrapper[4862]: I1128 11:35:41.885664 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4" Nov 28 11:35:42 crc kubenswrapper[4862]: I1128 11:35:42.219752 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q8zgx" podStartSLOduration=2.572545881 podStartE2EDuration="5.219724205s" podCreationTimestamp="2025-11-28 11:35:37 +0000 UTC" firstStartedPulling="2025-11-28 11:35:38.833354265 +0000 UTC m=+816.365868196" lastFinishedPulling="2025-11-28 11:35:41.480532559 +0000 UTC m=+819.013046520" observedRunningTime="2025-11-28 11:35:41.910065304 +0000 UTC m=+819.442579295" watchObservedRunningTime="2025-11-28 11:35:42.219724205 +0000 UTC m=+819.752238166" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.004959 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9"] Nov 28 11:35:47 crc kubenswrapper[4862]: E1128 11:35:47.005566 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7347f1d-97ee-4391-9d84-08348032071c" containerName="pull" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.005583 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7347f1d-97ee-4391-9d84-08348032071c" containerName="pull" Nov 28 11:35:47 crc kubenswrapper[4862]: E1128 11:35:47.005605 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7347f1d-97ee-4391-9d84-08348032071c" containerName="extract" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.005613 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7347f1d-97ee-4391-9d84-08348032071c" containerName="extract" Nov 28 11:35:47 crc kubenswrapper[4862]: E1128 11:35:47.005626 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7347f1d-97ee-4391-9d84-08348032071c" containerName="util" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.005633 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7347f1d-97ee-4391-9d84-08348032071c" containerName="util" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.005742 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7347f1d-97ee-4391-9d84-08348032071c" containerName="extract" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.006209 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.010193 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.010212 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.010268 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-dvplx" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.015392 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9"] Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.133374 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzqz6\" (UniqueName: \"kubernetes.io/projected/d26ccddf-e0a8-474b-87e3-2c9437e8f2eb-kube-api-access-gzqz6\") pod \"nmstate-operator-5b5b58f5c8-jlnt9\" (UID: \"d26ccddf-e0a8-474b-87e3-2c9437e8f2eb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.234195 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzqz6\" (UniqueName: \"kubernetes.io/projected/d26ccddf-e0a8-474b-87e3-2c9437e8f2eb-kube-api-access-gzqz6\") pod \"nmstate-operator-5b5b58f5c8-jlnt9\" (UID: \"d26ccddf-e0a8-474b-87e3-2c9437e8f2eb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.261002 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzqz6\" (UniqueName: \"kubernetes.io/projected/d26ccddf-e0a8-474b-87e3-2c9437e8f2eb-kube-api-access-gzqz6\") pod \"nmstate-operator-5b5b58f5c8-jlnt9\" (UID: \"d26ccddf-e0a8-474b-87e3-2c9437e8f2eb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.323964 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9" Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.542482 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9"] Nov 28 11:35:47 crc kubenswrapper[4862]: I1128 11:35:47.920942 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9" event={"ID":"d26ccddf-e0a8-474b-87e3-2c9437e8f2eb","Type":"ContainerStarted","Data":"96ef6f31c08e8fff342f9a7e087aeb2719c5bfc93cfba94817c1d2e839fc7c3b"} Nov 28 11:35:48 crc kubenswrapper[4862]: I1128 11:35:48.038733 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:48 crc kubenswrapper[4862]: I1128 11:35:48.038830 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:49 crc kubenswrapper[4862]: I1128 11:35:49.106208 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q8zgx" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" containerName="registry-server" probeResult="failure" output=< Nov 28 11:35:49 crc kubenswrapper[4862]: timeout: failed to connect service ":50051" within 1s Nov 28 11:35:49 crc kubenswrapper[4862]: > Nov 28 11:35:54 crc kubenswrapper[4862]: I1128 11:35:54.987730 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9" event={"ID":"d26ccddf-e0a8-474b-87e3-2c9437e8f2eb","Type":"ContainerStarted","Data":"39d566a25e19d1327d017bbc9ccefe518b5b2abb8d886405ef16a2d2a8090fcf"} Nov 28 11:35:55 crc kubenswrapper[4862]: I1128 11:35:55.018559 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-jlnt9" podStartSLOduration=3.097246236 podStartE2EDuration="9.018522339s" podCreationTimestamp="2025-11-28 11:35:46 +0000 UTC" firstStartedPulling="2025-11-28 11:35:47.550081907 +0000 UTC m=+825.082595838" lastFinishedPulling="2025-11-28 11:35:53.47135801 +0000 UTC m=+831.003871941" observedRunningTime="2025-11-28 11:35:55.010716209 +0000 UTC m=+832.543230170" watchObservedRunningTime="2025-11-28 11:35:55.018522339 +0000 UTC m=+832.551036310" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.043601 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc"] Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.046970 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.055505 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-pn9k8" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.071446 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc"] Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.079578 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb"] Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.081321 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.087544 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.096571 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-gn2tn"] Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.103425 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.126896 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb"] Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.202729 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6l79\" (UniqueName: \"kubernetes.io/projected/ac91e27c-fbee-4e6a-9ae1-001d68ada16d-kube-api-access-v6l79\") pod \"nmstate-webhook-5f6d4c5ccb-vtpdb\" (UID: \"ac91e27c-fbee-4e6a-9ae1-001d68ada16d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.203023 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac91e27c-fbee-4e6a-9ae1-001d68ada16d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-vtpdb\" (UID: \"ac91e27c-fbee-4e6a-9ae1-001d68ada16d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.203073 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-nmstate-lock\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.203104 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-dbus-socket\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.203124 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lsx6\" (UniqueName: \"kubernetes.io/projected/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-kube-api-access-8lsx6\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.203143 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f699x\" (UniqueName: \"kubernetes.io/projected/bf1dc471-6ca1-44fb-9c66-d905ad48ac57-kube-api-access-f699x\") pod \"nmstate-metrics-7f946cbc9-pngdc\" (UID: \"bf1dc471-6ca1-44fb-9c66-d905ad48ac57\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.203202 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-ovs-socket\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.272737 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk"] Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.273414 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.275346 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-mcb8v" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.275409 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.275418 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.284392 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk"] Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.303816 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f699x\" (UniqueName: \"kubernetes.io/projected/bf1dc471-6ca1-44fb-9c66-d905ad48ac57-kube-api-access-f699x\") pod \"nmstate-metrics-7f946cbc9-pngdc\" (UID: \"bf1dc471-6ca1-44fb-9c66-d905ad48ac57\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.303868 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-ovs-socket\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.303906 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6l79\" (UniqueName: \"kubernetes.io/projected/ac91e27c-fbee-4e6a-9ae1-001d68ada16d-kube-api-access-v6l79\") pod \"nmstate-webhook-5f6d4c5ccb-vtpdb\" (UID: \"ac91e27c-fbee-4e6a-9ae1-001d68ada16d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.303935 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac91e27c-fbee-4e6a-9ae1-001d68ada16d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-vtpdb\" (UID: \"ac91e27c-fbee-4e6a-9ae1-001d68ada16d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.303952 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-nmstate-lock\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.303975 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-dbus-socket\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.303993 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lsx6\" (UniqueName: \"kubernetes.io/projected/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-kube-api-access-8lsx6\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.304233 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-ovs-socket\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.304242 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-nmstate-lock\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.304472 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-dbus-socket\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.311466 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac91e27c-fbee-4e6a-9ae1-001d68ada16d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-vtpdb\" (UID: \"ac91e27c-fbee-4e6a-9ae1-001d68ada16d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.322495 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lsx6\" (UniqueName: \"kubernetes.io/projected/3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11-kube-api-access-8lsx6\") pod \"nmstate-handler-gn2tn\" (UID: \"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11\") " pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.325185 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6l79\" (UniqueName: \"kubernetes.io/projected/ac91e27c-fbee-4e6a-9ae1-001d68ada16d-kube-api-access-v6l79\") pod \"nmstate-webhook-5f6d4c5ccb-vtpdb\" (UID: \"ac91e27c-fbee-4e6a-9ae1-001d68ada16d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.325689 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f699x\" (UniqueName: \"kubernetes.io/projected/bf1dc471-6ca1-44fb-9c66-d905ad48ac57-kube-api-access-f699x\") pod \"nmstate-metrics-7f946cbc9-pngdc\" (UID: \"bf1dc471-6ca1-44fb-9c66-d905ad48ac57\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.375358 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.406043 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/49d902d7-ba96-4136-ab9b-e544620bb88a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-k65jk\" (UID: \"49d902d7-ba96-4136-ab9b-e544620bb88a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.406163 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/49d902d7-ba96-4136-ab9b-e544620bb88a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-k65jk\" (UID: \"49d902d7-ba96-4136-ab9b-e544620bb88a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.406203 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lpkl\" (UniqueName: \"kubernetes.io/projected/49d902d7-ba96-4136-ab9b-e544620bb88a-kube-api-access-8lpkl\") pod \"nmstate-console-plugin-7fbb5f6569-k65jk\" (UID: \"49d902d7-ba96-4136-ab9b-e544620bb88a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.408740 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.428618 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:35:56 crc kubenswrapper[4862]: W1128 11:35:56.464879 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ff44fb9_0a14_4ec9_a5bf_0410a47d9a11.slice/crio-260e3d78975e1e28472f39da90f7701656ab8c694de183af036cf94ac074e78b WatchSource:0}: Error finding container 260e3d78975e1e28472f39da90f7701656ab8c694de183af036cf94ac074e78b: Status 404 returned error can't find the container with id 260e3d78975e1e28472f39da90f7701656ab8c694de183af036cf94ac074e78b Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.507898 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lpkl\" (UniqueName: \"kubernetes.io/projected/49d902d7-ba96-4136-ab9b-e544620bb88a-kube-api-access-8lpkl\") pod \"nmstate-console-plugin-7fbb5f6569-k65jk\" (UID: \"49d902d7-ba96-4136-ab9b-e544620bb88a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.507986 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/49d902d7-ba96-4136-ab9b-e544620bb88a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-k65jk\" (UID: \"49d902d7-ba96-4136-ab9b-e544620bb88a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.508016 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/49d902d7-ba96-4136-ab9b-e544620bb88a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-k65jk\" (UID: \"49d902d7-ba96-4136-ab9b-e544620bb88a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.508849 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/49d902d7-ba96-4136-ab9b-e544620bb88a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-k65jk\" (UID: \"49d902d7-ba96-4136-ab9b-e544620bb88a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.516063 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/49d902d7-ba96-4136-ab9b-e544620bb88a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-k65jk\" (UID: \"49d902d7-ba96-4136-ab9b-e544620bb88a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.516227 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-76f59b595d-rhghc"] Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.516827 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.535404 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lpkl\" (UniqueName: \"kubernetes.io/projected/49d902d7-ba96-4136-ab9b-e544620bb88a-kube-api-access-8lpkl\") pod \"nmstate-console-plugin-7fbb5f6569-k65jk\" (UID: \"49d902d7-ba96-4136-ab9b-e544620bb88a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.535839 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76f59b595d-rhghc"] Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.587574 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.610543 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8qjz\" (UniqueName: \"kubernetes.io/projected/de655aba-8e50-4b09-b1ea-a08edf71e348-kube-api-access-f8qjz\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.610763 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-oauth-serving-cert\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.610790 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/de655aba-8e50-4b09-b1ea-a08edf71e348-console-serving-cert\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.610814 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-console-config\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.610837 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-trusted-ca-bundle\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.610874 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-service-ca\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.610930 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/de655aba-8e50-4b09-b1ea-a08edf71e348-console-oauth-config\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.711862 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/de655aba-8e50-4b09-b1ea-a08edf71e348-console-oauth-config\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.711907 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8qjz\" (UniqueName: \"kubernetes.io/projected/de655aba-8e50-4b09-b1ea-a08edf71e348-kube-api-access-f8qjz\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.711925 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-oauth-serving-cert\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.711945 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/de655aba-8e50-4b09-b1ea-a08edf71e348-console-serving-cert\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.711965 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-console-config\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.711985 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-trusted-ca-bundle\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.712014 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-service-ca\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.713380 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-console-config\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.716715 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-service-ca\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.719876 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-oauth-serving-cert\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.720059 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de655aba-8e50-4b09-b1ea-a08edf71e348-trusted-ca-bundle\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.721492 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/de655aba-8e50-4b09-b1ea-a08edf71e348-console-oauth-config\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.723943 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/de655aba-8e50-4b09-b1ea-a08edf71e348-console-serving-cert\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.726703 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8qjz\" (UniqueName: \"kubernetes.io/projected/de655aba-8e50-4b09-b1ea-a08edf71e348-kube-api-access-f8qjz\") pod \"console-76f59b595d-rhghc\" (UID: \"de655aba-8e50-4b09-b1ea-a08edf71e348\") " pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.829675 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc"] Nov 28 11:35:56 crc kubenswrapper[4862]: W1128 11:35:56.833254 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf1dc471_6ca1_44fb_9c66_d905ad48ac57.slice/crio-2b9a00bf23b4f4b2a5cb8ebf7282a685420fbccf5d17e140a169177a3efe54a2 WatchSource:0}: Error finding container 2b9a00bf23b4f4b2a5cb8ebf7282a685420fbccf5d17e140a169177a3efe54a2: Status 404 returned error can't find the container with id 2b9a00bf23b4f4b2a5cb8ebf7282a685420fbccf5d17e140a169177a3efe54a2 Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.837632 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.848555 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb"] Nov 28 11:35:56 crc kubenswrapper[4862]: W1128 11:35:56.850697 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac91e27c_fbee_4e6a_9ae1_001d68ada16d.slice/crio-f76b11caaeab429c1082b03930d97b8df012a0213cb76213ed0f612d2e6deb11 WatchSource:0}: Error finding container f76b11caaeab429c1082b03930d97b8df012a0213cb76213ed0f612d2e6deb11: Status 404 returned error can't find the container with id f76b11caaeab429c1082b03930d97b8df012a0213cb76213ed0f612d2e6deb11 Nov 28 11:35:56 crc kubenswrapper[4862]: I1128 11:35:56.893144 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk"] Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.009653 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gn2tn" event={"ID":"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11","Type":"ContainerStarted","Data":"260e3d78975e1e28472f39da90f7701656ab8c694de183af036cf94ac074e78b"} Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.021672 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" event={"ID":"ac91e27c-fbee-4e6a-9ae1-001d68ada16d","Type":"ContainerStarted","Data":"f76b11caaeab429c1082b03930d97b8df012a0213cb76213ed0f612d2e6deb11"} Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.022962 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" event={"ID":"49d902d7-ba96-4136-ab9b-e544620bb88a","Type":"ContainerStarted","Data":"da4bc757c3d43a6c7db6546c3fe32938c81d9af220c22fdd0573f1b4fb2f2f45"} Nov 28 11:35:57 crc kubenswrapper[4862]: W1128 11:35:57.024518 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde655aba_8e50_4b09_b1ea_a08edf71e348.slice/crio-f0e44bc3f52d824a90629ca7ad852fb3e108f6086bb5416d0bbc423b32141a57 WatchSource:0}: Error finding container f0e44bc3f52d824a90629ca7ad852fb3e108f6086bb5416d0bbc423b32141a57: Status 404 returned error can't find the container with id f0e44bc3f52d824a90629ca7ad852fb3e108f6086bb5416d0bbc423b32141a57 Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.024628 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc" event={"ID":"bf1dc471-6ca1-44fb-9c66-d905ad48ac57","Type":"ContainerStarted","Data":"2b9a00bf23b4f4b2a5cb8ebf7282a685420fbccf5d17e140a169177a3efe54a2"} Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.030497 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76f59b595d-rhghc"] Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.276439 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h56rf"] Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.280466 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.292030 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h56rf"] Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.428806 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-catalog-content\") pod \"certified-operators-h56rf\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.429582 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wxgx\" (UniqueName: \"kubernetes.io/projected/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-kube-api-access-8wxgx\") pod \"certified-operators-h56rf\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.429661 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-utilities\") pod \"certified-operators-h56rf\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.531956 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-catalog-content\") pod \"certified-operators-h56rf\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.532258 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wxgx\" (UniqueName: \"kubernetes.io/projected/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-kube-api-access-8wxgx\") pod \"certified-operators-h56rf\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.532366 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-utilities\") pod \"certified-operators-h56rf\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.532615 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-catalog-content\") pod \"certified-operators-h56rf\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.532934 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-utilities\") pod \"certified-operators-h56rf\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.569455 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wxgx\" (UniqueName: \"kubernetes.io/projected/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-kube-api-access-8wxgx\") pod \"certified-operators-h56rf\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:57 crc kubenswrapper[4862]: I1128 11:35:57.696667 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:35:58 crc kubenswrapper[4862]: I1128 11:35:58.061640 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76f59b595d-rhghc" event={"ID":"de655aba-8e50-4b09-b1ea-a08edf71e348","Type":"ContainerStarted","Data":"3480e6302bab3855b0a761548607e15182885e95a922258dd1d3952e331003cb"} Nov 28 11:35:58 crc kubenswrapper[4862]: I1128 11:35:58.062044 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76f59b595d-rhghc" event={"ID":"de655aba-8e50-4b09-b1ea-a08edf71e348","Type":"ContainerStarted","Data":"f0e44bc3f52d824a90629ca7ad852fb3e108f6086bb5416d0bbc423b32141a57"} Nov 28 11:35:58 crc kubenswrapper[4862]: I1128 11:35:58.080451 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-76f59b595d-rhghc" podStartSLOduration=2.080433772 podStartE2EDuration="2.080433772s" podCreationTimestamp="2025-11-28 11:35:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:35:58.078842537 +0000 UTC m=+835.611356468" watchObservedRunningTime="2025-11-28 11:35:58.080433772 +0000 UTC m=+835.612947713" Nov 28 11:35:58 crc kubenswrapper[4862]: I1128 11:35:58.111375 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:58 crc kubenswrapper[4862]: I1128 11:35:58.115729 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h56rf"] Nov 28 11:35:58 crc kubenswrapper[4862]: W1128 11:35:58.146346 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dcf0eeb_c27d_4f17_97f3_d4f73f95601b.slice/crio-bcbe341def7eca72e90d3bdd40bb096231b3a1377446c3ac4f56f3b3aebe7e41 WatchSource:0}: Error finding container bcbe341def7eca72e90d3bdd40bb096231b3a1377446c3ac4f56f3b3aebe7e41: Status 404 returned error can't find the container with id bcbe341def7eca72e90d3bdd40bb096231b3a1377446c3ac4f56f3b3aebe7e41 Nov 28 11:35:58 crc kubenswrapper[4862]: I1128 11:35:58.157532 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:35:59 crc kubenswrapper[4862]: I1128 11:35:59.073337 4862 generic.go:334] "Generic (PLEG): container finished" podID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerID="93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6" exitCode=0 Nov 28 11:35:59 crc kubenswrapper[4862]: I1128 11:35:59.074313 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h56rf" event={"ID":"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b","Type":"ContainerDied","Data":"93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6"} Nov 28 11:35:59 crc kubenswrapper[4862]: I1128 11:35:59.074375 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h56rf" event={"ID":"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b","Type":"ContainerStarted","Data":"bcbe341def7eca72e90d3bdd40bb096231b3a1377446c3ac4f56f3b3aebe7e41"} Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.057038 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8zgx"] Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.057964 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q8zgx" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" containerName="registry-server" containerID="cri-o://7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c" gracePeriod=2 Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.096207 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h56rf" event={"ID":"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b","Type":"ContainerStarted","Data":"2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd"} Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.097960 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gn2tn" event={"ID":"3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11","Type":"ContainerStarted","Data":"31a34296e1d077844a5192882bd280c5866d6565f110474952a98f0823f02819"} Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.098136 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.101758 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" event={"ID":"ac91e27c-fbee-4e6a-9ae1-001d68ada16d","Type":"ContainerStarted","Data":"ed94e386f5a0efc1fac4a48f37d26e03de3e5a3bddc410a94aa594e63789dcee"} Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.101839 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.110723 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc" event={"ID":"bf1dc471-6ca1-44fb-9c66-d905ad48ac57","Type":"ContainerStarted","Data":"f1435465db55cbcf27aa6289cb85a6317bf8359c4bc670748c0706d295d4185d"} Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.112539 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" event={"ID":"49d902d7-ba96-4136-ab9b-e544620bb88a","Type":"ContainerStarted","Data":"a61f42f3bb08c0ecc8e805b38818bcec048f40352ba32c8136893abe0d8a2967"} Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.135345 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-gn2tn" podStartSLOduration=1.153430268 podStartE2EDuration="5.135318057s" podCreationTimestamp="2025-11-28 11:35:56 +0000 UTC" firstStartedPulling="2025-11-28 11:35:56.468144661 +0000 UTC m=+834.000658582" lastFinishedPulling="2025-11-28 11:36:00.45003244 +0000 UTC m=+837.982546371" observedRunningTime="2025-11-28 11:36:01.1286573 +0000 UTC m=+838.661171221" watchObservedRunningTime="2025-11-28 11:36:01.135318057 +0000 UTC m=+838.667831978" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.185127 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k65jk" podStartSLOduration=1.671804285 podStartE2EDuration="5.185081229s" podCreationTimestamp="2025-11-28 11:35:56 +0000 UTC" firstStartedPulling="2025-11-28 11:35:56.913145341 +0000 UTC m=+834.445659262" lastFinishedPulling="2025-11-28 11:36:00.426422275 +0000 UTC m=+837.958936206" observedRunningTime="2025-11-28 11:36:01.150225717 +0000 UTC m=+838.682739678" watchObservedRunningTime="2025-11-28 11:36:01.185081229 +0000 UTC m=+838.717595150" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.403897 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.424542 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" podStartSLOduration=1.8416891789999998 podStartE2EDuration="5.424522791s" podCreationTimestamp="2025-11-28 11:35:56 +0000 UTC" firstStartedPulling="2025-11-28 11:35:56.85344091 +0000 UTC m=+834.385954831" lastFinishedPulling="2025-11-28 11:36:00.436274512 +0000 UTC m=+837.968788443" observedRunningTime="2025-11-28 11:36:01.179914463 +0000 UTC m=+838.712428394" watchObservedRunningTime="2025-11-28 11:36:01.424522791 +0000 UTC m=+838.957036722" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.499337 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5txv\" (UniqueName: \"kubernetes.io/projected/bfa1d663-78ab-4131-b299-b93f4108245f-kube-api-access-x5txv\") pod \"bfa1d663-78ab-4131-b299-b93f4108245f\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.499507 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-utilities\") pod \"bfa1d663-78ab-4131-b299-b93f4108245f\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.499554 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-catalog-content\") pod \"bfa1d663-78ab-4131-b299-b93f4108245f\" (UID: \"bfa1d663-78ab-4131-b299-b93f4108245f\") " Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.500607 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-utilities" (OuterVolumeSpecName: "utilities") pod "bfa1d663-78ab-4131-b299-b93f4108245f" (UID: "bfa1d663-78ab-4131-b299-b93f4108245f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.505908 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfa1d663-78ab-4131-b299-b93f4108245f-kube-api-access-x5txv" (OuterVolumeSpecName: "kube-api-access-x5txv") pod "bfa1d663-78ab-4131-b299-b93f4108245f" (UID: "bfa1d663-78ab-4131-b299-b93f4108245f"). InnerVolumeSpecName "kube-api-access-x5txv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.602228 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.602301 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5txv\" (UniqueName: \"kubernetes.io/projected/bfa1d663-78ab-4131-b299-b93f4108245f-kube-api-access-x5txv\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.623479 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfa1d663-78ab-4131-b299-b93f4108245f" (UID: "bfa1d663-78ab-4131-b299-b93f4108245f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:36:01 crc kubenswrapper[4862]: I1128 11:36:01.704430 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa1d663-78ab-4131-b299-b93f4108245f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.121900 4862 generic.go:334] "Generic (PLEG): container finished" podID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerID="2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd" exitCode=0 Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.122021 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h56rf" event={"ID":"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b","Type":"ContainerDied","Data":"2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd"} Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.124220 4862 generic.go:334] "Generic (PLEG): container finished" podID="bfa1d663-78ab-4131-b299-b93f4108245f" containerID="7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c" exitCode=0 Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.124431 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zgx" event={"ID":"bfa1d663-78ab-4131-b299-b93f4108245f","Type":"ContainerDied","Data":"7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c"} Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.124466 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zgx" event={"ID":"bfa1d663-78ab-4131-b299-b93f4108245f","Type":"ContainerDied","Data":"14a4a5fe7d7e237ec1bd446d04bae1af8494f0d327bd59b3afd5bb9be066f0dc"} Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.124491 4862 scope.go:117] "RemoveContainer" containerID="7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.124638 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8zgx" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.164707 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8zgx"] Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.167984 4862 scope.go:117] "RemoveContainer" containerID="ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.168696 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q8zgx"] Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.193515 4862 scope.go:117] "RemoveContainer" containerID="aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.209070 4862 scope.go:117] "RemoveContainer" containerID="7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c" Nov 28 11:36:02 crc kubenswrapper[4862]: E1128 11:36:02.209791 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c\": container with ID starting with 7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c not found: ID does not exist" containerID="7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.209841 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c"} err="failed to get container status \"7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c\": rpc error: code = NotFound desc = could not find container \"7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c\": container with ID starting with 7c38f640c23e91f7e1343dbc5ad2b27710e1fdef97633ddf6c1334c3771c417c not found: ID does not exist" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.209879 4862 scope.go:117] "RemoveContainer" containerID="ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e" Nov 28 11:36:02 crc kubenswrapper[4862]: E1128 11:36:02.210226 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e\": container with ID starting with ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e not found: ID does not exist" containerID="ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.210255 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e"} err="failed to get container status \"ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e\": rpc error: code = NotFound desc = could not find container \"ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e\": container with ID starting with ba87ae2d770c4d085912f4425c577888afb655f6e25de420ab4194faf6fbf74e not found: ID does not exist" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.210283 4862 scope.go:117] "RemoveContainer" containerID="aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab" Nov 28 11:36:02 crc kubenswrapper[4862]: E1128 11:36:02.210699 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab\": container with ID starting with aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab not found: ID does not exist" containerID="aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.210728 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab"} err="failed to get container status \"aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab\": rpc error: code = NotFound desc = could not find container \"aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab\": container with ID starting with aebab76541a30f3dabeec8128cd011f8f9767d024159efeb14d37e4e4b3e74ab not found: ID does not exist" Nov 28 11:36:02 crc kubenswrapper[4862]: I1128 11:36:02.878613 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" path="/var/lib/kubelet/pods/bfa1d663-78ab-4131-b299-b93f4108245f/volumes" Nov 28 11:36:03 crc kubenswrapper[4862]: I1128 11:36:03.149382 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h56rf" event={"ID":"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b","Type":"ContainerStarted","Data":"4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575"} Nov 28 11:36:03 crc kubenswrapper[4862]: I1128 11:36:03.168188 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc" event={"ID":"bf1dc471-6ca1-44fb-9c66-d905ad48ac57","Type":"ContainerStarted","Data":"c338bc84298876856859bbf3d35fe878326ac13a1cda56c441351d9ac74a38fb"} Nov 28 11:36:03 crc kubenswrapper[4862]: I1128 11:36:03.214082 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h56rf" podStartSLOduration=2.812671313 podStartE2EDuration="6.214049835s" podCreationTimestamp="2025-11-28 11:35:57 +0000 UTC" firstStartedPulling="2025-11-28 11:35:59.244475522 +0000 UTC m=+836.776989443" lastFinishedPulling="2025-11-28 11:36:02.645854014 +0000 UTC m=+840.178367965" observedRunningTime="2025-11-28 11:36:03.191026447 +0000 UTC m=+840.723540408" watchObservedRunningTime="2025-11-28 11:36:03.214049835 +0000 UTC m=+840.746563776" Nov 28 11:36:03 crc kubenswrapper[4862]: I1128 11:36:03.214483 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pngdc" podStartSLOduration=1.4036063730000001 podStartE2EDuration="7.214477877s" podCreationTimestamp="2025-11-28 11:35:56 +0000 UTC" firstStartedPulling="2025-11-28 11:35:56.83569311 +0000 UTC m=+834.368207031" lastFinishedPulling="2025-11-28 11:36:02.646564614 +0000 UTC m=+840.179078535" observedRunningTime="2025-11-28 11:36:03.208083417 +0000 UTC m=+840.740597358" watchObservedRunningTime="2025-11-28 11:36:03.214477877 +0000 UTC m=+840.746991818" Nov 28 11:36:06 crc kubenswrapper[4862]: I1128 11:36:06.469984 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-gn2tn" Nov 28 11:36:06 crc kubenswrapper[4862]: I1128 11:36:06.851455 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:36:06 crc kubenswrapper[4862]: I1128 11:36:06.851775 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:36:06 crc kubenswrapper[4862]: I1128 11:36:06.851991 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:36:06 crc kubenswrapper[4862]: I1128 11:36:06.858644 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-76f59b595d-rhghc" Nov 28 11:36:06 crc kubenswrapper[4862]: I1128 11:36:06.979128 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pzshw"] Nov 28 11:36:07 crc kubenswrapper[4862]: I1128 11:36:07.697147 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:36:07 crc kubenswrapper[4862]: I1128 11:36:07.697246 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:36:07 crc kubenswrapper[4862]: I1128 11:36:07.770480 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:36:08 crc kubenswrapper[4862]: I1128 11:36:08.271079 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:36:08 crc kubenswrapper[4862]: I1128 11:36:08.291592 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:36:08 crc kubenswrapper[4862]: I1128 11:36:08.291670 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:36:08 crc kubenswrapper[4862]: I1128 11:36:08.291731 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:36:08 crc kubenswrapper[4862]: I1128 11:36:08.292713 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cd460d7f02e8c5ea1bc5f7ee547d7f0b295f608f496d14e974dc72d5aea4821d"} pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:36:08 crc kubenswrapper[4862]: I1128 11:36:08.292810 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" containerID="cri-o://cd460d7f02e8c5ea1bc5f7ee547d7f0b295f608f496d14e974dc72d5aea4821d" gracePeriod=600 Nov 28 11:36:08 crc kubenswrapper[4862]: I1128 11:36:08.340236 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h56rf"] Nov 28 11:36:10 crc kubenswrapper[4862]: I1128 11:36:10.219045 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6845439-1a4c-4f28-9ba4-29149d545995" containerID="cd460d7f02e8c5ea1bc5f7ee547d7f0b295f608f496d14e974dc72d5aea4821d" exitCode=0 Nov 28 11:36:10 crc kubenswrapper[4862]: I1128 11:36:10.219174 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerDied","Data":"cd460d7f02e8c5ea1bc5f7ee547d7f0b295f608f496d14e974dc72d5aea4821d"} Nov 28 11:36:10 crc kubenswrapper[4862]: I1128 11:36:10.219279 4862 scope.go:117] "RemoveContainer" containerID="6d25f125eb927ad9f5d5ebf41c0bfc64bfcc8ea0c1567eba44d6b656d68eb6c7" Nov 28 11:36:10 crc kubenswrapper[4862]: I1128 11:36:10.219444 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h56rf" podUID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerName="registry-server" containerID="cri-o://4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575" gracePeriod=2 Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.163300 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.232144 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"5e63a1b14a90646d7f5b8ec0b596a3abf9ade9420df0f277e56e83cc3a51ca76"} Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.233711 4862 generic.go:334] "Generic (PLEG): container finished" podID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerID="4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575" exitCode=0 Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.233767 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h56rf" event={"ID":"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b","Type":"ContainerDied","Data":"4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575"} Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.233822 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h56rf" event={"ID":"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b","Type":"ContainerDied","Data":"bcbe341def7eca72e90d3bdd40bb096231b3a1377446c3ac4f56f3b3aebe7e41"} Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.233849 4862 scope.go:117] "RemoveContainer" containerID="4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.233916 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h56rf" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.260758 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-utilities\") pod \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.260966 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wxgx\" (UniqueName: \"kubernetes.io/projected/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-kube-api-access-8wxgx\") pod \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.261006 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-catalog-content\") pod \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\" (UID: \"9dcf0eeb-c27d-4f17-97f3-d4f73f95601b\") " Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.262802 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-utilities" (OuterVolumeSpecName: "utilities") pod "9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" (UID: "9dcf0eeb-c27d-4f17-97f3-d4f73f95601b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.265686 4862 scope.go:117] "RemoveContainer" containerID="2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.272942 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-kube-api-access-8wxgx" (OuterVolumeSpecName: "kube-api-access-8wxgx") pod "9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" (UID: "9dcf0eeb-c27d-4f17-97f3-d4f73f95601b"). InnerVolumeSpecName "kube-api-access-8wxgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.312074 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" (UID: "9dcf0eeb-c27d-4f17-97f3-d4f73f95601b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.314819 4862 scope.go:117] "RemoveContainer" containerID="93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.334828 4862 scope.go:117] "RemoveContainer" containerID="4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575" Nov 28 11:36:11 crc kubenswrapper[4862]: E1128 11:36:11.335335 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575\": container with ID starting with 4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575 not found: ID does not exist" containerID="4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.335387 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575"} err="failed to get container status \"4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575\": rpc error: code = NotFound desc = could not find container \"4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575\": container with ID starting with 4e0c1eba35fa144923b644d6d270da6cb309240fdfb9ccdfeb59b179be9c9575 not found: ID does not exist" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.335424 4862 scope.go:117] "RemoveContainer" containerID="2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd" Nov 28 11:36:11 crc kubenswrapper[4862]: E1128 11:36:11.335741 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd\": container with ID starting with 2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd not found: ID does not exist" containerID="2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.335787 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd"} err="failed to get container status \"2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd\": rpc error: code = NotFound desc = could not find container \"2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd\": container with ID starting with 2f292a055be37c7d4d6e77c14811c31970585088d86d2bf1d2f770423bbd78fd not found: ID does not exist" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.335814 4862 scope.go:117] "RemoveContainer" containerID="93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6" Nov 28 11:36:11 crc kubenswrapper[4862]: E1128 11:36:11.336349 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6\": container with ID starting with 93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6 not found: ID does not exist" containerID="93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.336390 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6"} err="failed to get container status \"93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6\": rpc error: code = NotFound desc = could not find container \"93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6\": container with ID starting with 93262ecfba0190dfc644f2c2b37fa4c9ab3c9ab3deb91d87fd5a1d2eec6b12f6 not found: ID does not exist" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.362340 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.362379 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.362392 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wxgx\" (UniqueName: \"kubernetes.io/projected/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b-kube-api-access-8wxgx\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.578648 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h56rf"] Nov 28 11:36:11 crc kubenswrapper[4862]: I1128 11:36:11.586579 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h56rf"] Nov 28 11:36:12 crc kubenswrapper[4862]: I1128 11:36:12.874285 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" path="/var/lib/kubelet/pods/9dcf0eeb-c27d-4f17-97f3-d4f73f95601b/volumes" Nov 28 11:36:16 crc kubenswrapper[4862]: I1128 11:36:16.418906 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vtpdb" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.696488 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p"] Nov 28 11:36:31 crc kubenswrapper[4862]: E1128 11:36:31.697584 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" containerName="extract-utilities" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.697600 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" containerName="extract-utilities" Nov 28 11:36:31 crc kubenswrapper[4862]: E1128 11:36:31.697612 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerName="registry-server" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.697618 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerName="registry-server" Nov 28 11:36:31 crc kubenswrapper[4862]: E1128 11:36:31.697630 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" containerName="extract-content" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.697639 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" containerName="extract-content" Nov 28 11:36:31 crc kubenswrapper[4862]: E1128 11:36:31.697655 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerName="extract-utilities" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.697661 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerName="extract-utilities" Nov 28 11:36:31 crc kubenswrapper[4862]: E1128 11:36:31.697669 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerName="extract-content" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.697675 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerName="extract-content" Nov 28 11:36:31 crc kubenswrapper[4862]: E1128 11:36:31.697686 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" containerName="registry-server" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.697692 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" containerName="registry-server" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.697792 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfa1d663-78ab-4131-b299-b93f4108245f" containerName="registry-server" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.697808 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dcf0eeb-c27d-4f17-97f3-d4f73f95601b" containerName="registry-server" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.698748 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.701115 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.709021 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p"] Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.809812 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.809893 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.810030 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7d7k\" (UniqueName: \"kubernetes.io/projected/9726dcde-7041-402a-9ab8-9a7b04e3bff6-kube-api-access-z7d7k\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.911261 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.911378 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7d7k\" (UniqueName: \"kubernetes.io/projected/9726dcde-7041-402a-9ab8-9a7b04e3bff6-kube-api-access-z7d7k\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.911435 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.911944 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.912198 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:31 crc kubenswrapper[4862]: I1128 11:36:31.942503 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7d7k\" (UniqueName: \"kubernetes.io/projected/9726dcde-7041-402a-9ab8-9a7b04e3bff6-kube-api-access-z7d7k\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.019268 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.048968 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-pzshw" podUID="b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" containerName="console" containerID="cri-o://34f3e7a06e9df9bdfc47fd83388550a63944b3f82f342cda02971eda7c96f379" gracePeriod=15 Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.399840 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pzshw_b0a9a925-c659-48aa-a4dd-6d002b9ad2d7/console/0.log" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.400475 4862 generic.go:334] "Generic (PLEG): container finished" podID="b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" containerID="34f3e7a06e9df9bdfc47fd83388550a63944b3f82f342cda02971eda7c96f379" exitCode=2 Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.400525 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pzshw" event={"ID":"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7","Type":"ContainerDied","Data":"34f3e7a06e9df9bdfc47fd83388550a63944b3f82f342cda02971eda7c96f379"} Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.499538 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pzshw_b0a9a925-c659-48aa-a4dd-6d002b9ad2d7/console/0.log" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.499652 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.541080 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p"] Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.621834 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-trusted-ca-bundle\") pod \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.621977 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-config\") pod \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.623127 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" (UID: "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.624921 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-config" (OuterVolumeSpecName: "console-config") pod "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" (UID: "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.625746 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-serving-cert\") pod \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.627752 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-oauth-config\") pod \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.628501 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" (UID: "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.628929 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-oauth-serving-cert\") pod \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.629160 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8fwm\" (UniqueName: \"kubernetes.io/projected/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-kube-api-access-k8fwm\") pod \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.629323 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-service-ca\") pod \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\" (UID: \"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7\") " Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.629911 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-service-ca" (OuterVolumeSpecName: "service-ca") pod "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" (UID: "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.631360 4862 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.631406 4862 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.631425 4862 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.631443 4862 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.635155 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-kube-api-access-k8fwm" (OuterVolumeSpecName: "kube-api-access-k8fwm") pod "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" (UID: "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7"). InnerVolumeSpecName "kube-api-access-k8fwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.635452 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" (UID: "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.636462 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" (UID: "b0a9a925-c659-48aa-a4dd-6d002b9ad2d7"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.732784 4862 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.732842 4862 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:32 crc kubenswrapper[4862]: I1128 11:36:32.732862 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8fwm\" (UniqueName: \"kubernetes.io/projected/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7-kube-api-access-k8fwm\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:33 crc kubenswrapper[4862]: I1128 11:36:33.408977 4862 generic.go:334] "Generic (PLEG): container finished" podID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerID="487d7d879a22954f13fcd585f3813e9b4a33990dbe7ae11b09246f2a29afe642" exitCode=0 Nov 28 11:36:33 crc kubenswrapper[4862]: I1128 11:36:33.409059 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" event={"ID":"9726dcde-7041-402a-9ab8-9a7b04e3bff6","Type":"ContainerDied","Data":"487d7d879a22954f13fcd585f3813e9b4a33990dbe7ae11b09246f2a29afe642"} Nov 28 11:36:33 crc kubenswrapper[4862]: I1128 11:36:33.409151 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" event={"ID":"9726dcde-7041-402a-9ab8-9a7b04e3bff6","Type":"ContainerStarted","Data":"9ae09d1e7d2ecbd5035b487bca3d1ec6b8ea45e6ad810476f8dff8575da47d37"} Nov 28 11:36:33 crc kubenswrapper[4862]: I1128 11:36:33.410965 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pzshw_b0a9a925-c659-48aa-a4dd-6d002b9ad2d7/console/0.log" Nov 28 11:36:33 crc kubenswrapper[4862]: I1128 11:36:33.411009 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pzshw" event={"ID":"b0a9a925-c659-48aa-a4dd-6d002b9ad2d7","Type":"ContainerDied","Data":"3db17258d2ca71f27c91c0077b3e217702126185806a5630b324097cee3146d9"} Nov 28 11:36:33 crc kubenswrapper[4862]: I1128 11:36:33.411055 4862 scope.go:117] "RemoveContainer" containerID="34f3e7a06e9df9bdfc47fd83388550a63944b3f82f342cda02971eda7c96f379" Nov 28 11:36:33 crc kubenswrapper[4862]: I1128 11:36:33.411190 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pzshw" Nov 28 11:36:33 crc kubenswrapper[4862]: I1128 11:36:33.447083 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pzshw"] Nov 28 11:36:33 crc kubenswrapper[4862]: I1128 11:36:33.450852 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-pzshw"] Nov 28 11:36:34 crc kubenswrapper[4862]: I1128 11:36:34.854405 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" path="/var/lib/kubelet/pods/b0a9a925-c659-48aa-a4dd-6d002b9ad2d7/volumes" Nov 28 11:36:35 crc kubenswrapper[4862]: I1128 11:36:35.440714 4862 generic.go:334] "Generic (PLEG): container finished" podID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerID="74ac91528738e22fab8d73f005b73657f9afa885f55d095ad15bd0dc6f90ce08" exitCode=0 Nov 28 11:36:35 crc kubenswrapper[4862]: I1128 11:36:35.440798 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" event={"ID":"9726dcde-7041-402a-9ab8-9a7b04e3bff6","Type":"ContainerDied","Data":"74ac91528738e22fab8d73f005b73657f9afa885f55d095ad15bd0dc6f90ce08"} Nov 28 11:36:36 crc kubenswrapper[4862]: I1128 11:36:36.452698 4862 generic.go:334] "Generic (PLEG): container finished" podID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerID="3fbfcf85fb6e40bbfcfb4df6d7dc5b450936b1bce41931ceb80a9e7c067334d5" exitCode=0 Nov 28 11:36:36 crc kubenswrapper[4862]: I1128 11:36:36.452771 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" event={"ID":"9726dcde-7041-402a-9ab8-9a7b04e3bff6","Type":"ContainerDied","Data":"3fbfcf85fb6e40bbfcfb4df6d7dc5b450936b1bce41931ceb80a9e7c067334d5"} Nov 28 11:36:37 crc kubenswrapper[4862]: I1128 11:36:37.764453 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:37 crc kubenswrapper[4862]: I1128 11:36:37.916081 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7d7k\" (UniqueName: \"kubernetes.io/projected/9726dcde-7041-402a-9ab8-9a7b04e3bff6-kube-api-access-z7d7k\") pod \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " Nov 28 11:36:37 crc kubenswrapper[4862]: I1128 11:36:37.916352 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-util\") pod \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " Nov 28 11:36:37 crc kubenswrapper[4862]: I1128 11:36:37.916426 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-bundle\") pod \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\" (UID: \"9726dcde-7041-402a-9ab8-9a7b04e3bff6\") " Nov 28 11:36:37 crc kubenswrapper[4862]: I1128 11:36:37.918413 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-bundle" (OuterVolumeSpecName: "bundle") pod "9726dcde-7041-402a-9ab8-9a7b04e3bff6" (UID: "9726dcde-7041-402a-9ab8-9a7b04e3bff6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:36:37 crc kubenswrapper[4862]: I1128 11:36:37.927706 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9726dcde-7041-402a-9ab8-9a7b04e3bff6-kube-api-access-z7d7k" (OuterVolumeSpecName: "kube-api-access-z7d7k") pod "9726dcde-7041-402a-9ab8-9a7b04e3bff6" (UID: "9726dcde-7041-402a-9ab8-9a7b04e3bff6"). InnerVolumeSpecName "kube-api-access-z7d7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:36:37 crc kubenswrapper[4862]: I1128 11:36:37.937865 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-util" (OuterVolumeSpecName: "util") pod "9726dcde-7041-402a-9ab8-9a7b04e3bff6" (UID: "9726dcde-7041-402a-9ab8-9a7b04e3bff6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:36:38 crc kubenswrapper[4862]: I1128 11:36:38.017850 4862 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-util\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:38 crc kubenswrapper[4862]: I1128 11:36:38.017898 4862 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9726dcde-7041-402a-9ab8-9a7b04e3bff6-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:38 crc kubenswrapper[4862]: I1128 11:36:38.017920 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7d7k\" (UniqueName: \"kubernetes.io/projected/9726dcde-7041-402a-9ab8-9a7b04e3bff6-kube-api-access-z7d7k\") on node \"crc\" DevicePath \"\"" Nov 28 11:36:38 crc kubenswrapper[4862]: I1128 11:36:38.475595 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" event={"ID":"9726dcde-7041-402a-9ab8-9a7b04e3bff6","Type":"ContainerDied","Data":"9ae09d1e7d2ecbd5035b487bca3d1ec6b8ea45e6ad810476f8dff8575da47d37"} Nov 28 11:36:38 crc kubenswrapper[4862]: I1128 11:36:38.475665 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ae09d1e7d2ecbd5035b487bca3d1ec6b8ea45e6ad810476f8dff8575da47d37" Nov 28 11:36:38 crc kubenswrapper[4862]: I1128 11:36:38.475717 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.792273 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp"] Nov 28 11:36:46 crc kubenswrapper[4862]: E1128 11:36:46.792907 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerName="pull" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.792919 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerName="pull" Nov 28 11:36:46 crc kubenswrapper[4862]: E1128 11:36:46.792927 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerName="extract" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.792933 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerName="extract" Nov 28 11:36:46 crc kubenswrapper[4862]: E1128 11:36:46.792946 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerName="util" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.792953 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerName="util" Nov 28 11:36:46 crc kubenswrapper[4862]: E1128 11:36:46.792962 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" containerName="console" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.792967 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" containerName="console" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.793056 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9726dcde-7041-402a-9ab8-9a7b04e3bff6" containerName="extract" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.793065 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0a9a925-c659-48aa-a4dd-6d002b9ad2d7" containerName="console" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.793528 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.796391 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.796770 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.798056 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.800573 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.800814 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-7cwhl" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.816217 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp"] Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.969818 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aeebe377-e303-4234-9fcc-c43d7d22f738-apiservice-cert\") pod \"metallb-operator-controller-manager-78fccdcc4f-7d2dp\" (UID: \"aeebe377-e303-4234-9fcc-c43d7d22f738\") " pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.970494 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d622r\" (UniqueName: \"kubernetes.io/projected/aeebe377-e303-4234-9fcc-c43d7d22f738-kube-api-access-d622r\") pod \"metallb-operator-controller-manager-78fccdcc4f-7d2dp\" (UID: \"aeebe377-e303-4234-9fcc-c43d7d22f738\") " pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:46 crc kubenswrapper[4862]: I1128 11:36:46.970576 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aeebe377-e303-4234-9fcc-c43d7d22f738-webhook-cert\") pod \"metallb-operator-controller-manager-78fccdcc4f-7d2dp\" (UID: \"aeebe377-e303-4234-9fcc-c43d7d22f738\") " pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.071711 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aeebe377-e303-4234-9fcc-c43d7d22f738-apiservice-cert\") pod \"metallb-operator-controller-manager-78fccdcc4f-7d2dp\" (UID: \"aeebe377-e303-4234-9fcc-c43d7d22f738\") " pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.071813 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d622r\" (UniqueName: \"kubernetes.io/projected/aeebe377-e303-4234-9fcc-c43d7d22f738-kube-api-access-d622r\") pod \"metallb-operator-controller-manager-78fccdcc4f-7d2dp\" (UID: \"aeebe377-e303-4234-9fcc-c43d7d22f738\") " pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.071843 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aeebe377-e303-4234-9fcc-c43d7d22f738-webhook-cert\") pod \"metallb-operator-controller-manager-78fccdcc4f-7d2dp\" (UID: \"aeebe377-e303-4234-9fcc-c43d7d22f738\") " pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.078666 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aeebe377-e303-4234-9fcc-c43d7d22f738-webhook-cert\") pod \"metallb-operator-controller-manager-78fccdcc4f-7d2dp\" (UID: \"aeebe377-e303-4234-9fcc-c43d7d22f738\") " pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.079114 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aeebe377-e303-4234-9fcc-c43d7d22f738-apiservice-cert\") pod \"metallb-operator-controller-manager-78fccdcc4f-7d2dp\" (UID: \"aeebe377-e303-4234-9fcc-c43d7d22f738\") " pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.094188 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d622r\" (UniqueName: \"kubernetes.io/projected/aeebe377-e303-4234-9fcc-c43d7d22f738-kube-api-access-d622r\") pod \"metallb-operator-controller-manager-78fccdcc4f-7d2dp\" (UID: \"aeebe377-e303-4234-9fcc-c43d7d22f738\") " pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.111048 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.164101 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s"] Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.164785 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.166604 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.166861 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-s6v9r" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.167005 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.173454 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7-apiservice-cert\") pod \"metallb-operator-webhook-server-f7f698684-bhd4s\" (UID: \"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7\") " pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.173500 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cbls\" (UniqueName: \"kubernetes.io/projected/9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7-kube-api-access-9cbls\") pod \"metallb-operator-webhook-server-f7f698684-bhd4s\" (UID: \"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7\") " pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.173741 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7-webhook-cert\") pod \"metallb-operator-webhook-server-f7f698684-bhd4s\" (UID: \"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7\") " pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.182842 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s"] Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.278642 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7-apiservice-cert\") pod \"metallb-operator-webhook-server-f7f698684-bhd4s\" (UID: \"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7\") " pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.278969 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cbls\" (UniqueName: \"kubernetes.io/projected/9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7-kube-api-access-9cbls\") pod \"metallb-operator-webhook-server-f7f698684-bhd4s\" (UID: \"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7\") " pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.279008 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7-webhook-cert\") pod \"metallb-operator-webhook-server-f7f698684-bhd4s\" (UID: \"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7\") " pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.284863 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7-apiservice-cert\") pod \"metallb-operator-webhook-server-f7f698684-bhd4s\" (UID: \"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7\") " pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.287589 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7-webhook-cert\") pod \"metallb-operator-webhook-server-f7f698684-bhd4s\" (UID: \"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7\") " pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.296452 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cbls\" (UniqueName: \"kubernetes.io/projected/9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7-kube-api-access-9cbls\") pod \"metallb-operator-webhook-server-f7f698684-bhd4s\" (UID: \"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7\") " pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.479420 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.605236 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp"] Nov 28 11:36:47 crc kubenswrapper[4862]: I1128 11:36:47.737976 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s"] Nov 28 11:36:47 crc kubenswrapper[4862]: W1128 11:36:47.746369 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fb4f150_6dfc_4aed_a8c6_3c0ba89d39d7.slice/crio-40935adfe675a8b8a9fc9787ee6f008c66fe3957077125d9f9d099b0ccd63a81 WatchSource:0}: Error finding container 40935adfe675a8b8a9fc9787ee6f008c66fe3957077125d9f9d099b0ccd63a81: Status 404 returned error can't find the container with id 40935adfe675a8b8a9fc9787ee6f008c66fe3957077125d9f9d099b0ccd63a81 Nov 28 11:36:48 crc kubenswrapper[4862]: I1128 11:36:48.545292 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" event={"ID":"aeebe377-e303-4234-9fcc-c43d7d22f738","Type":"ContainerStarted","Data":"668d3a115df38d89f4f3268a74d58284b9aaa625b95fd9797cafeb387d34fba3"} Nov 28 11:36:48 crc kubenswrapper[4862]: I1128 11:36:48.546684 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" event={"ID":"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7","Type":"ContainerStarted","Data":"40935adfe675a8b8a9fc9787ee6f008c66fe3957077125d9f9d099b0ccd63a81"} Nov 28 11:36:50 crc kubenswrapper[4862]: I1128 11:36:50.935389 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4ll77"] Nov 28 11:36:50 crc kubenswrapper[4862]: I1128 11:36:50.937063 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:50 crc kubenswrapper[4862]: I1128 11:36:50.951069 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4ll77"] Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.053786 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvzs6\" (UniqueName: \"kubernetes.io/projected/43acba18-3299-4423-8ff0-3f803294c5a8-kube-api-access-fvzs6\") pod \"redhat-marketplace-4ll77\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.053846 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-utilities\") pod \"redhat-marketplace-4ll77\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.053869 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-catalog-content\") pod \"redhat-marketplace-4ll77\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.154569 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvzs6\" (UniqueName: \"kubernetes.io/projected/43acba18-3299-4423-8ff0-3f803294c5a8-kube-api-access-fvzs6\") pod \"redhat-marketplace-4ll77\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.154633 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-utilities\") pod \"redhat-marketplace-4ll77\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.154653 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-catalog-content\") pod \"redhat-marketplace-4ll77\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.155163 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-catalog-content\") pod \"redhat-marketplace-4ll77\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.155378 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-utilities\") pod \"redhat-marketplace-4ll77\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.177806 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvzs6\" (UniqueName: \"kubernetes.io/projected/43acba18-3299-4423-8ff0-3f803294c5a8-kube-api-access-fvzs6\") pod \"redhat-marketplace-4ll77\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.276042 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.566062 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" event={"ID":"aeebe377-e303-4234-9fcc-c43d7d22f738","Type":"ContainerStarted","Data":"d68be10af687d531e151779c436bb40e5fb649c1f6b51fafe79944748e8e31cb"} Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.566278 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:36:51 crc kubenswrapper[4862]: I1128 11:36:51.594421 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" podStartSLOduration=2.379873607 podStartE2EDuration="5.59440264s" podCreationTimestamp="2025-11-28 11:36:46 +0000 UTC" firstStartedPulling="2025-11-28 11:36:47.636714574 +0000 UTC m=+885.169228495" lastFinishedPulling="2025-11-28 11:36:50.851243607 +0000 UTC m=+888.383757528" observedRunningTime="2025-11-28 11:36:51.590639941 +0000 UTC m=+889.123153872" watchObservedRunningTime="2025-11-28 11:36:51.59440264 +0000 UTC m=+889.126916561" Nov 28 11:36:52 crc kubenswrapper[4862]: I1128 11:36:52.964489 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4ll77"] Nov 28 11:36:53 crc kubenswrapper[4862]: I1128 11:36:53.578028 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" event={"ID":"9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7","Type":"ContainerStarted","Data":"ba90136c76b927737f73ce091db4b76165866c3486b75b3fee3cb8ede7a0d9d1"} Nov 28 11:36:53 crc kubenswrapper[4862]: I1128 11:36:53.578412 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:36:53 crc kubenswrapper[4862]: I1128 11:36:53.580879 4862 generic.go:334] "Generic (PLEG): container finished" podID="43acba18-3299-4423-8ff0-3f803294c5a8" containerID="453f6ee96be10f8b8070606793f2c36252c4760f10f7424bedf7f804648e4bed" exitCode=0 Nov 28 11:36:53 crc kubenswrapper[4862]: I1128 11:36:53.580958 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4ll77" event={"ID":"43acba18-3299-4423-8ff0-3f803294c5a8","Type":"ContainerDied","Data":"453f6ee96be10f8b8070606793f2c36252c4760f10f7424bedf7f804648e4bed"} Nov 28 11:36:53 crc kubenswrapper[4862]: I1128 11:36:53.581000 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4ll77" event={"ID":"43acba18-3299-4423-8ff0-3f803294c5a8","Type":"ContainerStarted","Data":"5316d7b984bddacdf78b025383253d822ebb3affe07d55aa0ad2753adca03b90"} Nov 28 11:36:53 crc kubenswrapper[4862]: I1128 11:36:53.606197 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" podStartSLOduration=1.574061948 podStartE2EDuration="6.606169624s" podCreationTimestamp="2025-11-28 11:36:47 +0000 UTC" firstStartedPulling="2025-11-28 11:36:47.748583935 +0000 UTC m=+885.281097856" lastFinishedPulling="2025-11-28 11:36:52.780691601 +0000 UTC m=+890.313205532" observedRunningTime="2025-11-28 11:36:53.597507961 +0000 UTC m=+891.130021902" watchObservedRunningTime="2025-11-28 11:36:53.606169624 +0000 UTC m=+891.138683555" Nov 28 11:36:54 crc kubenswrapper[4862]: I1128 11:36:54.589236 4862 generic.go:334] "Generic (PLEG): container finished" podID="43acba18-3299-4423-8ff0-3f803294c5a8" containerID="c4fad9c7c622bca656678e4f3a4045b8afae9c7e3b7f9f29f7863d99facfdea0" exitCode=0 Nov 28 11:36:54 crc kubenswrapper[4862]: I1128 11:36:54.589369 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4ll77" event={"ID":"43acba18-3299-4423-8ff0-3f803294c5a8","Type":"ContainerDied","Data":"c4fad9c7c622bca656678e4f3a4045b8afae9c7e3b7f9f29f7863d99facfdea0"} Nov 28 11:36:56 crc kubenswrapper[4862]: I1128 11:36:56.606286 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4ll77" event={"ID":"43acba18-3299-4423-8ff0-3f803294c5a8","Type":"ContainerStarted","Data":"7392e660f2be974d2c5370014fa280bc4148c6ba666cea81bade27dfb9393649"} Nov 28 11:36:56 crc kubenswrapper[4862]: I1128 11:36:56.647374 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4ll77" podStartSLOduration=4.550644485 podStartE2EDuration="6.647345914s" podCreationTimestamp="2025-11-28 11:36:50 +0000 UTC" firstStartedPulling="2025-11-28 11:36:53.582585646 +0000 UTC m=+891.115099567" lastFinishedPulling="2025-11-28 11:36:55.679287065 +0000 UTC m=+893.211800996" observedRunningTime="2025-11-28 11:36:56.640419602 +0000 UTC m=+894.172933523" watchObservedRunningTime="2025-11-28 11:36:56.647345914 +0000 UTC m=+894.179859855" Nov 28 11:37:01 crc kubenswrapper[4862]: I1128 11:37:01.277063 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:37:01 crc kubenswrapper[4862]: I1128 11:37:01.277489 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:37:01 crc kubenswrapper[4862]: I1128 11:37:01.330982 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:37:01 crc kubenswrapper[4862]: I1128 11:37:01.686189 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:37:03 crc kubenswrapper[4862]: I1128 11:37:03.648075 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4ll77"] Nov 28 11:37:03 crc kubenswrapper[4862]: I1128 11:37:03.651029 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4ll77" podUID="43acba18-3299-4423-8ff0-3f803294c5a8" containerName="registry-server" containerID="cri-o://7392e660f2be974d2c5370014fa280bc4148c6ba666cea81bade27dfb9393649" gracePeriod=2 Nov 28 11:37:05 crc kubenswrapper[4862]: I1128 11:37:05.675916 4862 generic.go:334] "Generic (PLEG): container finished" podID="43acba18-3299-4423-8ff0-3f803294c5a8" containerID="7392e660f2be974d2c5370014fa280bc4148c6ba666cea81bade27dfb9393649" exitCode=0 Nov 28 11:37:05 crc kubenswrapper[4862]: I1128 11:37:05.675985 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4ll77" event={"ID":"43acba18-3299-4423-8ff0-3f803294c5a8","Type":"ContainerDied","Data":"7392e660f2be974d2c5370014fa280bc4148c6ba666cea81bade27dfb9393649"} Nov 28 11:37:05 crc kubenswrapper[4862]: I1128 11:37:05.928229 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.059732 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-utilities\") pod \"43acba18-3299-4423-8ff0-3f803294c5a8\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.059918 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvzs6\" (UniqueName: \"kubernetes.io/projected/43acba18-3299-4423-8ff0-3f803294c5a8-kube-api-access-fvzs6\") pod \"43acba18-3299-4423-8ff0-3f803294c5a8\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.060206 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-catalog-content\") pod \"43acba18-3299-4423-8ff0-3f803294c5a8\" (UID: \"43acba18-3299-4423-8ff0-3f803294c5a8\") " Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.061843 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-utilities" (OuterVolumeSpecName: "utilities") pod "43acba18-3299-4423-8ff0-3f803294c5a8" (UID: "43acba18-3299-4423-8ff0-3f803294c5a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.069082 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43acba18-3299-4423-8ff0-3f803294c5a8-kube-api-access-fvzs6" (OuterVolumeSpecName: "kube-api-access-fvzs6") pod "43acba18-3299-4423-8ff0-3f803294c5a8" (UID: "43acba18-3299-4423-8ff0-3f803294c5a8"). InnerVolumeSpecName "kube-api-access-fvzs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.078915 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43acba18-3299-4423-8ff0-3f803294c5a8" (UID: "43acba18-3299-4423-8ff0-3f803294c5a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.161790 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvzs6\" (UniqueName: \"kubernetes.io/projected/43acba18-3299-4423-8ff0-3f803294c5a8-kube-api-access-fvzs6\") on node \"crc\" DevicePath \"\"" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.161830 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.161846 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43acba18-3299-4423-8ff0-3f803294c5a8-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.683204 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4ll77" event={"ID":"43acba18-3299-4423-8ff0-3f803294c5a8","Type":"ContainerDied","Data":"5316d7b984bddacdf78b025383253d822ebb3affe07d55aa0ad2753adca03b90"} Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.683270 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4ll77" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.683293 4862 scope.go:117] "RemoveContainer" containerID="7392e660f2be974d2c5370014fa280bc4148c6ba666cea81bade27dfb9393649" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.708325 4862 scope.go:117] "RemoveContainer" containerID="c4fad9c7c622bca656678e4f3a4045b8afae9c7e3b7f9f29f7863d99facfdea0" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.722065 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4ll77"] Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.729159 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4ll77"] Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.760307 4862 scope.go:117] "RemoveContainer" containerID="453f6ee96be10f8b8070606793f2c36252c4760f10f7424bedf7f804648e4bed" Nov 28 11:37:06 crc kubenswrapper[4862]: I1128 11:37:06.861301 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43acba18-3299-4423-8ff0-3f803294c5a8" path="/var/lib/kubelet/pods/43acba18-3299-4423-8ff0-3f803294c5a8/volumes" Nov 28 11:37:07 crc kubenswrapper[4862]: I1128 11:37:07.491083 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-f7f698684-bhd4s" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.259949 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fks5f"] Nov 28 11:37:11 crc kubenswrapper[4862]: E1128 11:37:11.260622 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43acba18-3299-4423-8ff0-3f803294c5a8" containerName="registry-server" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.260643 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="43acba18-3299-4423-8ff0-3f803294c5a8" containerName="registry-server" Nov 28 11:37:11 crc kubenswrapper[4862]: E1128 11:37:11.260669 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43acba18-3299-4423-8ff0-3f803294c5a8" containerName="extract-utilities" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.260682 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="43acba18-3299-4423-8ff0-3f803294c5a8" containerName="extract-utilities" Nov 28 11:37:11 crc kubenswrapper[4862]: E1128 11:37:11.260702 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43acba18-3299-4423-8ff0-3f803294c5a8" containerName="extract-content" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.260714 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="43acba18-3299-4423-8ff0-3f803294c5a8" containerName="extract-content" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.260910 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="43acba18-3299-4423-8ff0-3f803294c5a8" containerName="registry-server" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.262207 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.281522 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fks5f"] Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.436479 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-utilities\") pod \"community-operators-fks5f\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.436523 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hfg\" (UniqueName: \"kubernetes.io/projected/6fb03cda-378e-477e-89a0-46e283d59299-kube-api-access-87hfg\") pod \"community-operators-fks5f\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.436677 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-catalog-content\") pod \"community-operators-fks5f\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.537692 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-utilities\") pod \"community-operators-fks5f\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.537743 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hfg\" (UniqueName: \"kubernetes.io/projected/6fb03cda-378e-477e-89a0-46e283d59299-kube-api-access-87hfg\") pod \"community-operators-fks5f\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.537774 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-catalog-content\") pod \"community-operators-fks5f\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.538326 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-catalog-content\") pod \"community-operators-fks5f\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.538329 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-utilities\") pod \"community-operators-fks5f\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.562011 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hfg\" (UniqueName: \"kubernetes.io/projected/6fb03cda-378e-477e-89a0-46e283d59299-kube-api-access-87hfg\") pod \"community-operators-fks5f\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:11 crc kubenswrapper[4862]: I1128 11:37:11.606767 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:12 crc kubenswrapper[4862]: I1128 11:37:12.061128 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fks5f"] Nov 28 11:37:12 crc kubenswrapper[4862]: I1128 11:37:12.743967 4862 generic.go:334] "Generic (PLEG): container finished" podID="6fb03cda-378e-477e-89a0-46e283d59299" containerID="0b3b4171d13146dcda1834d902d2c4588c9a363030be9323713b298138c15e53" exitCode=0 Nov 28 11:37:12 crc kubenswrapper[4862]: I1128 11:37:12.744129 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fks5f" event={"ID":"6fb03cda-378e-477e-89a0-46e283d59299","Type":"ContainerDied","Data":"0b3b4171d13146dcda1834d902d2c4588c9a363030be9323713b298138c15e53"} Nov 28 11:37:12 crc kubenswrapper[4862]: I1128 11:37:12.744439 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fks5f" event={"ID":"6fb03cda-378e-477e-89a0-46e283d59299","Type":"ContainerStarted","Data":"5f790e37f184e7a4c438dc6411ccc93fd9039e6381e78ee05bf84a797d2894f6"} Nov 28 11:37:13 crc kubenswrapper[4862]: I1128 11:37:13.754274 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fks5f" event={"ID":"6fb03cda-378e-477e-89a0-46e283d59299","Type":"ContainerStarted","Data":"b93bed0632402eb8c07cdaebc9495ad1ee842e31bb626e142d9a0607bff3fec5"} Nov 28 11:37:14 crc kubenswrapper[4862]: I1128 11:37:14.766577 4862 generic.go:334] "Generic (PLEG): container finished" podID="6fb03cda-378e-477e-89a0-46e283d59299" containerID="b93bed0632402eb8c07cdaebc9495ad1ee842e31bb626e142d9a0607bff3fec5" exitCode=0 Nov 28 11:37:14 crc kubenswrapper[4862]: I1128 11:37:14.766687 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fks5f" event={"ID":"6fb03cda-378e-477e-89a0-46e283d59299","Type":"ContainerDied","Data":"b93bed0632402eb8c07cdaebc9495ad1ee842e31bb626e142d9a0607bff3fec5"} Nov 28 11:37:16 crc kubenswrapper[4862]: I1128 11:37:16.785174 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fks5f" event={"ID":"6fb03cda-378e-477e-89a0-46e283d59299","Type":"ContainerStarted","Data":"ca6798caababb158955abb2446cd1c3c07e73e557bcaaf4e729c14448f47d816"} Nov 28 11:37:16 crc kubenswrapper[4862]: I1128 11:37:16.803564 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fks5f" podStartSLOduration=2.838161847 podStartE2EDuration="5.803546236s" podCreationTimestamp="2025-11-28 11:37:11 +0000 UTC" firstStartedPulling="2025-11-28 11:37:12.746786762 +0000 UTC m=+910.279300693" lastFinishedPulling="2025-11-28 11:37:15.712171121 +0000 UTC m=+913.244685082" observedRunningTime="2025-11-28 11:37:16.800042685 +0000 UTC m=+914.332556656" watchObservedRunningTime="2025-11-28 11:37:16.803546236 +0000 UTC m=+914.336060167" Nov 28 11:37:21 crc kubenswrapper[4862]: I1128 11:37:21.607850 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:21 crc kubenswrapper[4862]: I1128 11:37:21.608185 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:21 crc kubenswrapper[4862]: I1128 11:37:21.686084 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:21 crc kubenswrapper[4862]: I1128 11:37:21.901206 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:24 crc kubenswrapper[4862]: I1128 11:37:24.646012 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fks5f"] Nov 28 11:37:24 crc kubenswrapper[4862]: I1128 11:37:24.646656 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fks5f" podUID="6fb03cda-378e-477e-89a0-46e283d59299" containerName="registry-server" containerID="cri-o://ca6798caababb158955abb2446cd1c3c07e73e557bcaaf4e729c14448f47d816" gracePeriod=2 Nov 28 11:37:25 crc kubenswrapper[4862]: I1128 11:37:25.862149 4862 generic.go:334] "Generic (PLEG): container finished" podID="6fb03cda-378e-477e-89a0-46e283d59299" containerID="ca6798caababb158955abb2446cd1c3c07e73e557bcaaf4e729c14448f47d816" exitCode=0 Nov 28 11:37:25 crc kubenswrapper[4862]: I1128 11:37:25.862204 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fks5f" event={"ID":"6fb03cda-378e-477e-89a0-46e283d59299","Type":"ContainerDied","Data":"ca6798caababb158955abb2446cd1c3c07e73e557bcaaf4e729c14448f47d816"} Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.208439 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.355670 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87hfg\" (UniqueName: \"kubernetes.io/projected/6fb03cda-378e-477e-89a0-46e283d59299-kube-api-access-87hfg\") pod \"6fb03cda-378e-477e-89a0-46e283d59299\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.355877 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-catalog-content\") pod \"6fb03cda-378e-477e-89a0-46e283d59299\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.355921 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-utilities\") pod \"6fb03cda-378e-477e-89a0-46e283d59299\" (UID: \"6fb03cda-378e-477e-89a0-46e283d59299\") " Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.356751 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-utilities" (OuterVolumeSpecName: "utilities") pod "6fb03cda-378e-477e-89a0-46e283d59299" (UID: "6fb03cda-378e-477e-89a0-46e283d59299"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.360948 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fb03cda-378e-477e-89a0-46e283d59299-kube-api-access-87hfg" (OuterVolumeSpecName: "kube-api-access-87hfg") pod "6fb03cda-378e-477e-89a0-46e283d59299" (UID: "6fb03cda-378e-477e-89a0-46e283d59299"). InnerVolumeSpecName "kube-api-access-87hfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.403307 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fb03cda-378e-477e-89a0-46e283d59299" (UID: "6fb03cda-378e-477e-89a0-46e283d59299"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.457555 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87hfg\" (UniqueName: \"kubernetes.io/projected/6fb03cda-378e-477e-89a0-46e283d59299-kube-api-access-87hfg\") on node \"crc\" DevicePath \"\"" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.457587 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.457596 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb03cda-378e-477e-89a0-46e283d59299-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.872493 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fks5f" event={"ID":"6fb03cda-378e-477e-89a0-46e283d59299","Type":"ContainerDied","Data":"5f790e37f184e7a4c438dc6411ccc93fd9039e6381e78ee05bf84a797d2894f6"} Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.872865 4862 scope.go:117] "RemoveContainer" containerID="ca6798caababb158955abb2446cd1c3c07e73e557bcaaf4e729c14448f47d816" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.872587 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fks5f" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.899701 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fks5f"] Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.902040 4862 scope.go:117] "RemoveContainer" containerID="b93bed0632402eb8c07cdaebc9495ad1ee842e31bb626e142d9a0607bff3fec5" Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.905153 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fks5f"] Nov 28 11:37:26 crc kubenswrapper[4862]: I1128 11:37:26.920956 4862 scope.go:117] "RemoveContainer" containerID="0b3b4171d13146dcda1834d902d2c4588c9a363030be9323713b298138c15e53" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.114580 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-78fccdcc4f-7d2dp" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.936715 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv"] Nov 28 11:37:27 crc kubenswrapper[4862]: E1128 11:37:27.936964 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb03cda-378e-477e-89a0-46e283d59299" containerName="extract-utilities" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.936976 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb03cda-378e-477e-89a0-46e283d59299" containerName="extract-utilities" Nov 28 11:37:27 crc kubenswrapper[4862]: E1128 11:37:27.936989 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb03cda-378e-477e-89a0-46e283d59299" containerName="registry-server" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.936996 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb03cda-378e-477e-89a0-46e283d59299" containerName="registry-server" Nov 28 11:37:27 crc kubenswrapper[4862]: E1128 11:37:27.937015 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb03cda-378e-477e-89a0-46e283d59299" containerName="extract-content" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.937021 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb03cda-378e-477e-89a0-46e283d59299" containerName="extract-content" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.937137 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fb03cda-378e-477e-89a0-46e283d59299" containerName="registry-server" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.937544 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.940043 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-tz4mw"] Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.942063 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.946254 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.946723 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.946734 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-sdpnk" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.947034 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 28 11:37:27 crc kubenswrapper[4862]: I1128 11:37:27.958294 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv"] Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.078662 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qchmm"] Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079329 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtnkq\" (UniqueName: \"kubernetes.io/projected/01f38525-c12e-4706-b45c-5cba52dd6d02-kube-api-access-gtnkq\") pod \"frr-k8s-webhook-server-7fcb986d4-7w8vv\" (UID: \"01f38525-c12e-4706-b45c-5cba52dd6d02\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079387 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce782165-5659-4d92-b449-78a9cd04a9b2-metrics-certs\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079454 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-metrics\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079504 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-frr-conf\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079557 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-frr-sockets\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079582 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01f38525-c12e-4706-b45c-5cba52dd6d02-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7w8vv\" (UID: \"01f38525-c12e-4706-b45c-5cba52dd6d02\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079684 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtl9q\" (UniqueName: \"kubernetes.io/projected/ce782165-5659-4d92-b449-78a9cd04a9b2-kube-api-access-gtl9q\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079720 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-reloader\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079772 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.079780 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ce782165-5659-4d92-b449-78a9cd04a9b2-frr-startup\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.081024 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-xr594" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.081960 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.084406 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.084610 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.086505 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-c7z9m"] Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.087353 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.088758 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.111965 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-c7z9m"] Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.181805 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-memberlist\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.181872 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-frr-sockets\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.181905 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01f38525-c12e-4706-b45c-5cba52dd6d02-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7w8vv\" (UID: \"01f38525-c12e-4706-b45c-5cba52dd6d02\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.181946 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-metrics-certs\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.181980 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtl9q\" (UniqueName: \"kubernetes.io/projected/ce782165-5659-4d92-b449-78a9cd04a9b2-kube-api-access-gtl9q\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.182012 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-reloader\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.182036 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ce782165-5659-4d92-b449-78a9cd04a9b2-frr-startup\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.182076 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtnkq\" (UniqueName: \"kubernetes.io/projected/01f38525-c12e-4706-b45c-5cba52dd6d02-kube-api-access-gtnkq\") pod \"frr-k8s-webhook-server-7fcb986d4-7w8vv\" (UID: \"01f38525-c12e-4706-b45c-5cba52dd6d02\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.182133 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce782165-5659-4d92-b449-78a9cd04a9b2-metrics-certs\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.182163 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-metrics\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.182190 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-frr-conf\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.182214 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmsm8\" (UniqueName: \"kubernetes.io/projected/40c6570e-6b5e-481f-8b37-45e27aee43be-kube-api-access-bmsm8\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.182247 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/40c6570e-6b5e-481f-8b37-45e27aee43be-metallb-excludel2\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.182383 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-frr-sockets\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: E1128 11:37:28.182915 4862 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 28 11:37:28 crc kubenswrapper[4862]: E1128 11:37:28.182972 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce782165-5659-4d92-b449-78a9cd04a9b2-metrics-certs podName:ce782165-5659-4d92-b449-78a9cd04a9b2 nodeName:}" failed. No retries permitted until 2025-11-28 11:37:28.682957071 +0000 UTC m=+926.215470992 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ce782165-5659-4d92-b449-78a9cd04a9b2-metrics-certs") pod "frr-k8s-tz4mw" (UID: "ce782165-5659-4d92-b449-78a9cd04a9b2") : secret "frr-k8s-certs-secret" not found Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.183045 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-reloader\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.183153 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-metrics\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.183262 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ce782165-5659-4d92-b449-78a9cd04a9b2-frr-conf\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.183666 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ce782165-5659-4d92-b449-78a9cd04a9b2-frr-startup\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.189872 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01f38525-c12e-4706-b45c-5cba52dd6d02-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7w8vv\" (UID: \"01f38525-c12e-4706-b45c-5cba52dd6d02\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.203533 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtl9q\" (UniqueName: \"kubernetes.io/projected/ce782165-5659-4d92-b449-78a9cd04a9b2-kube-api-access-gtl9q\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.203632 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtnkq\" (UniqueName: \"kubernetes.io/projected/01f38525-c12e-4706-b45c-5cba52dd6d02-kube-api-access-gtnkq\") pod \"frr-k8s-webhook-server-7fcb986d4-7w8vv\" (UID: \"01f38525-c12e-4706-b45c-5cba52dd6d02\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.256271 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.283143 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/df58eb63-bc43-4d48-a2ce-d69a3bf053ef-metrics-certs\") pod \"controller-f8648f98b-c7z9m\" (UID: \"df58eb63-bc43-4d48-a2ce-d69a3bf053ef\") " pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.283240 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq9zk\" (UniqueName: \"kubernetes.io/projected/df58eb63-bc43-4d48-a2ce-d69a3bf053ef-kube-api-access-sq9zk\") pod \"controller-f8648f98b-c7z9m\" (UID: \"df58eb63-bc43-4d48-a2ce-d69a3bf053ef\") " pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.283281 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmsm8\" (UniqueName: \"kubernetes.io/projected/40c6570e-6b5e-481f-8b37-45e27aee43be-kube-api-access-bmsm8\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.283313 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/40c6570e-6b5e-481f-8b37-45e27aee43be-metallb-excludel2\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.283331 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-memberlist\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.283370 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df58eb63-bc43-4d48-a2ce-d69a3bf053ef-cert\") pod \"controller-f8648f98b-c7z9m\" (UID: \"df58eb63-bc43-4d48-a2ce-d69a3bf053ef\") " pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.283400 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-metrics-certs\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: E1128 11:37:28.283535 4862 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 28 11:37:28 crc kubenswrapper[4862]: E1128 11:37:28.283583 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-metrics-certs podName:40c6570e-6b5e-481f-8b37-45e27aee43be nodeName:}" failed. No retries permitted until 2025-11-28 11:37:28.78356758 +0000 UTC m=+926.316081491 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-metrics-certs") pod "speaker-qchmm" (UID: "40c6570e-6b5e-481f-8b37-45e27aee43be") : secret "speaker-certs-secret" not found Nov 28 11:37:28 crc kubenswrapper[4862]: E1128 11:37:28.283697 4862 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 11:37:28 crc kubenswrapper[4862]: E1128 11:37:28.283728 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-memberlist podName:40c6570e-6b5e-481f-8b37-45e27aee43be nodeName:}" failed. No retries permitted until 2025-11-28 11:37:28.783720594 +0000 UTC m=+926.316234515 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-memberlist") pod "speaker-qchmm" (UID: "40c6570e-6b5e-481f-8b37-45e27aee43be") : secret "metallb-memberlist" not found Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.284258 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/40c6570e-6b5e-481f-8b37-45e27aee43be-metallb-excludel2\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.304637 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmsm8\" (UniqueName: \"kubernetes.io/projected/40c6570e-6b5e-481f-8b37-45e27aee43be-kube-api-access-bmsm8\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.384830 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/df58eb63-bc43-4d48-a2ce-d69a3bf053ef-metrics-certs\") pod \"controller-f8648f98b-c7z9m\" (UID: \"df58eb63-bc43-4d48-a2ce-d69a3bf053ef\") " pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.384887 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq9zk\" (UniqueName: \"kubernetes.io/projected/df58eb63-bc43-4d48-a2ce-d69a3bf053ef-kube-api-access-sq9zk\") pod \"controller-f8648f98b-c7z9m\" (UID: \"df58eb63-bc43-4d48-a2ce-d69a3bf053ef\") " pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.384944 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df58eb63-bc43-4d48-a2ce-d69a3bf053ef-cert\") pod \"controller-f8648f98b-c7z9m\" (UID: \"df58eb63-bc43-4d48-a2ce-d69a3bf053ef\") " pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.388229 4862 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.388662 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/df58eb63-bc43-4d48-a2ce-d69a3bf053ef-metrics-certs\") pod \"controller-f8648f98b-c7z9m\" (UID: \"df58eb63-bc43-4d48-a2ce-d69a3bf053ef\") " pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.399545 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df58eb63-bc43-4d48-a2ce-d69a3bf053ef-cert\") pod \"controller-f8648f98b-c7z9m\" (UID: \"df58eb63-bc43-4d48-a2ce-d69a3bf053ef\") " pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.406886 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq9zk\" (UniqueName: \"kubernetes.io/projected/df58eb63-bc43-4d48-a2ce-d69a3bf053ef-kube-api-access-sq9zk\") pod \"controller-f8648f98b-c7z9m\" (UID: \"df58eb63-bc43-4d48-a2ce-d69a3bf053ef\") " pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.409427 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.587319 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-c7z9m"] Nov 28 11:37:28 crc kubenswrapper[4862]: W1128 11:37:28.593714 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf58eb63_bc43_4d48_a2ce_d69a3bf053ef.slice/crio-6ced36ff57a74eaba63036ef01dd42179cca4a7432cf32b135ed96f076f71e9c WatchSource:0}: Error finding container 6ced36ff57a74eaba63036ef01dd42179cca4a7432cf32b135ed96f076f71e9c: Status 404 returned error can't find the container with id 6ced36ff57a74eaba63036ef01dd42179cca4a7432cf32b135ed96f076f71e9c Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.647581 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv"] Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.688539 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce782165-5659-4d92-b449-78a9cd04a9b2-metrics-certs\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.692645 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce782165-5659-4d92-b449-78a9cd04a9b2-metrics-certs\") pod \"frr-k8s-tz4mw\" (UID: \"ce782165-5659-4d92-b449-78a9cd04a9b2\") " pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.790289 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-memberlist\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.790365 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-metrics-certs\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.794115 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-memberlist\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.794465 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40c6570e-6b5e-481f-8b37-45e27aee43be-metrics-certs\") pod \"speaker-qchmm\" (UID: \"40c6570e-6b5e-481f-8b37-45e27aee43be\") " pod="metallb-system/speaker-qchmm" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.847902 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fb03cda-378e-477e-89a0-46e283d59299" path="/var/lib/kubelet/pods/6fb03cda-378e-477e-89a0-46e283d59299/volumes" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.870669 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.887814 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-c7z9m" event={"ID":"df58eb63-bc43-4d48-a2ce-d69a3bf053ef","Type":"ContainerStarted","Data":"1530d5f458ca0c9e342917853c9cca79cfca623b67da3834f93c5c6505ca8dcd"} Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.887865 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-c7z9m" event={"ID":"df58eb63-bc43-4d48-a2ce-d69a3bf053ef","Type":"ContainerStarted","Data":"08dccf8d980f80e4263da8b47e59b6cb5eb99b814cfa0cca69a20da7f42818a1"} Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.887877 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-c7z9m" event={"ID":"df58eb63-bc43-4d48-a2ce-d69a3bf053ef","Type":"ContainerStarted","Data":"6ced36ff57a74eaba63036ef01dd42179cca4a7432cf32b135ed96f076f71e9c"} Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.887970 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.889701 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" event={"ID":"01f38525-c12e-4706-b45c-5cba52dd6d02","Type":"ContainerStarted","Data":"34aa9431457721259ff02ab90381d3d9902d13de4a4c3c0ba18e3a50f03975fe"} Nov 28 11:37:28 crc kubenswrapper[4862]: I1128 11:37:28.905854 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-c7z9m" podStartSLOduration=0.905828498 podStartE2EDuration="905.828498ms" podCreationTimestamp="2025-11-28 11:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:37:28.902249425 +0000 UTC m=+926.434763346" watchObservedRunningTime="2025-11-28 11:37:28.905828498 +0000 UTC m=+926.438342429" Nov 28 11:37:29 crc kubenswrapper[4862]: I1128 11:37:29.006615 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qchmm" Nov 28 11:37:29 crc kubenswrapper[4862]: W1128 11:37:29.029672 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40c6570e_6b5e_481f_8b37_45e27aee43be.slice/crio-47fa7a8475cd964a51010a9fd42fe3a3a6bb41c19a8df8666eaa0404ad0428cd WatchSource:0}: Error finding container 47fa7a8475cd964a51010a9fd42fe3a3a6bb41c19a8df8666eaa0404ad0428cd: Status 404 returned error can't find the container with id 47fa7a8475cd964a51010a9fd42fe3a3a6bb41c19a8df8666eaa0404ad0428cd Nov 28 11:37:29 crc kubenswrapper[4862]: I1128 11:37:29.899387 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerStarted","Data":"4b90e63531620e7dc3e91edf9b41c458dae9b442ea6490ecb2fd39b778b0fe91"} Nov 28 11:37:29 crc kubenswrapper[4862]: I1128 11:37:29.904274 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qchmm" event={"ID":"40c6570e-6b5e-481f-8b37-45e27aee43be","Type":"ContainerStarted","Data":"b5e35c5b11c16acf324af1b48ae57f6e6fee045f3c48b52defdba72d0f2f280d"} Nov 28 11:37:29 crc kubenswrapper[4862]: I1128 11:37:29.904345 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qchmm" event={"ID":"40c6570e-6b5e-481f-8b37-45e27aee43be","Type":"ContainerStarted","Data":"95983e3faf3ec5a0328459c7ee5b919c57f18917583d3de8b870166d9f387307"} Nov 28 11:37:29 crc kubenswrapper[4862]: I1128 11:37:29.904366 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qchmm" event={"ID":"40c6570e-6b5e-481f-8b37-45e27aee43be","Type":"ContainerStarted","Data":"47fa7a8475cd964a51010a9fd42fe3a3a6bb41c19a8df8666eaa0404ad0428cd"} Nov 28 11:37:29 crc kubenswrapper[4862]: I1128 11:37:29.904780 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qchmm" Nov 28 11:37:29 crc kubenswrapper[4862]: I1128 11:37:29.932356 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qchmm" podStartSLOduration=1.932331263 podStartE2EDuration="1.932331263s" podCreationTimestamp="2025-11-28 11:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:37:29.92978604 +0000 UTC m=+927.462300001" watchObservedRunningTime="2025-11-28 11:37:29.932331263 +0000 UTC m=+927.464845224" Nov 28 11:37:37 crc kubenswrapper[4862]: I1128 11:37:37.965403 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" event={"ID":"01f38525-c12e-4706-b45c-5cba52dd6d02","Type":"ContainerStarted","Data":"816f4777f28f0c4b0ed9a0cf87f7dd4e96c9f6330fec0c818118be3c4582ef26"} Nov 28 11:37:37 crc kubenswrapper[4862]: I1128 11:37:37.966035 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:37 crc kubenswrapper[4862]: I1128 11:37:37.969710 4862 generic.go:334] "Generic (PLEG): container finished" podID="ce782165-5659-4d92-b449-78a9cd04a9b2" containerID="e4c8f4b471cefce64eb8164ef9e247c49367f137e0dc779b91b6de109996a0aa" exitCode=0 Nov 28 11:37:37 crc kubenswrapper[4862]: I1128 11:37:37.969772 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerDied","Data":"e4c8f4b471cefce64eb8164ef9e247c49367f137e0dc779b91b6de109996a0aa"} Nov 28 11:37:38 crc kubenswrapper[4862]: I1128 11:37:38.001597 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" podStartSLOduration=2.897729466 podStartE2EDuration="11.001555874s" podCreationTimestamp="2025-11-28 11:37:27 +0000 UTC" firstStartedPulling="2025-11-28 11:37:28.657008169 +0000 UTC m=+926.189522080" lastFinishedPulling="2025-11-28 11:37:36.760834537 +0000 UTC m=+934.293348488" observedRunningTime="2025-11-28 11:37:37.991518724 +0000 UTC m=+935.524032695" watchObservedRunningTime="2025-11-28 11:37:38.001555874 +0000 UTC m=+935.534069825" Nov 28 11:37:38 crc kubenswrapper[4862]: I1128 11:37:38.414986 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-c7z9m" Nov 28 11:37:38 crc kubenswrapper[4862]: I1128 11:37:38.981924 4862 generic.go:334] "Generic (PLEG): container finished" podID="ce782165-5659-4d92-b449-78a9cd04a9b2" containerID="6cbe3ad4a1aa1e637f4d5a120f553fb223b4515bb1622d500881c15e7a4e9522" exitCode=0 Nov 28 11:37:38 crc kubenswrapper[4862]: I1128 11:37:38.981995 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerDied","Data":"6cbe3ad4a1aa1e637f4d5a120f553fb223b4515bb1622d500881c15e7a4e9522"} Nov 28 11:37:39 crc kubenswrapper[4862]: I1128 11:37:39.013080 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qchmm" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.698016 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq"] Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.699982 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.704888 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.713385 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq"] Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.804392 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg8vc\" (UniqueName: \"kubernetes.io/projected/95091706-4117-46dc-9405-f633f5467383-kube-api-access-tg8vc\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.804439 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.804479 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.905472 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg8vc\" (UniqueName: \"kubernetes.io/projected/95091706-4117-46dc-9405-f633f5467383-kube-api-access-tg8vc\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.905526 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.905567 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.906008 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.906375 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:40 crc kubenswrapper[4862]: I1128 11:37:40.942068 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg8vc\" (UniqueName: \"kubernetes.io/projected/95091706-4117-46dc-9405-f633f5467383-kube-api-access-tg8vc\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:41 crc kubenswrapper[4862]: I1128 11:37:41.003250 4862 generic.go:334] "Generic (PLEG): container finished" podID="ce782165-5659-4d92-b449-78a9cd04a9b2" containerID="bbe5f3c835cc6a151853976d70f1fa34f5dc42c3c9f5c224de32769b7c0795b0" exitCode=0 Nov 28 11:37:41 crc kubenswrapper[4862]: I1128 11:37:41.003302 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerDied","Data":"bbe5f3c835cc6a151853976d70f1fa34f5dc42c3c9f5c224de32769b7c0795b0"} Nov 28 11:37:41 crc kubenswrapper[4862]: I1128 11:37:41.032020 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:41 crc kubenswrapper[4862]: I1128 11:37:41.446778 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq"] Nov 28 11:37:41 crc kubenswrapper[4862]: W1128 11:37:41.454720 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95091706_4117_46dc_9405_f633f5467383.slice/crio-f06be3bf5921e56b72b37038fc1bb6d85143561777d7b2706df8c8e7c6e51514 WatchSource:0}: Error finding container f06be3bf5921e56b72b37038fc1bb6d85143561777d7b2706df8c8e7c6e51514: Status 404 returned error can't find the container with id f06be3bf5921e56b72b37038fc1bb6d85143561777d7b2706df8c8e7c6e51514 Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.009788 4862 generic.go:334] "Generic (PLEG): container finished" podID="95091706-4117-46dc-9405-f633f5467383" containerID="343b24529b0aeac41491566367676d5ed088920b12b793be923fd80ba74eedba" exitCode=0 Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.009868 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" event={"ID":"95091706-4117-46dc-9405-f633f5467383","Type":"ContainerDied","Data":"343b24529b0aeac41491566367676d5ed088920b12b793be923fd80ba74eedba"} Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.010202 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" event={"ID":"95091706-4117-46dc-9405-f633f5467383","Type":"ContainerStarted","Data":"f06be3bf5921e56b72b37038fc1bb6d85143561777d7b2706df8c8e7c6e51514"} Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.022565 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerStarted","Data":"16315ef1e5f2a1660f0dcfe0ad7958a30bd35e48de1ac7789238375776926a30"} Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.022604 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerStarted","Data":"9cf1ac6d76bb6964f67ba7a8f87156744e5556ce38ee5584b90a4e7367455cb2"} Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.022615 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerStarted","Data":"807dd8beaecf282212a0caa7980f331f75c923b5389ee17ffc21152ff1a0e021"} Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.022624 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerStarted","Data":"f31ee814699980146498108cb9a800e9ea2b5560997914c5ad60fac1cb34dbb4"} Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.022632 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerStarted","Data":"cee25aedffa7a5314dfdae4aa6b506be3cd9a37e98f93682be68094eb69ff2c6"} Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.022641 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tz4mw" event={"ID":"ce782165-5659-4d92-b449-78a9cd04a9b2","Type":"ContainerStarted","Data":"c3b07309bacb786d8dec8aad4b0729dc456543d4cc6963c3fbad1a9f71f6869d"} Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.022731 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:42 crc kubenswrapper[4862]: I1128 11:37:42.047736 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-tz4mw" podStartSLOduration=7.315980165 podStartE2EDuration="15.047716462s" podCreationTimestamp="2025-11-28 11:37:27 +0000 UTC" firstStartedPulling="2025-11-28 11:37:29.022572862 +0000 UTC m=+926.555086823" lastFinishedPulling="2025-11-28 11:37:36.754309199 +0000 UTC m=+934.286823120" observedRunningTime="2025-11-28 11:37:42.045778736 +0000 UTC m=+939.578292677" watchObservedRunningTime="2025-11-28 11:37:42.047716462 +0000 UTC m=+939.580230383" Nov 28 11:37:43 crc kubenswrapper[4862]: I1128 11:37:43.871502 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:43 crc kubenswrapper[4862]: I1128 11:37:43.911896 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:48 crc kubenswrapper[4862]: I1128 11:37:48.263827 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7w8vv" Nov 28 11:37:49 crc kubenswrapper[4862]: I1128 11:37:49.072179 4862 generic.go:334] "Generic (PLEG): container finished" podID="95091706-4117-46dc-9405-f633f5467383" containerID="a1d699e97f4f3066e110d2189b65713edeea31f22ce56dbff4967e5ead216561" exitCode=0 Nov 28 11:37:49 crc kubenswrapper[4862]: I1128 11:37:49.072242 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" event={"ID":"95091706-4117-46dc-9405-f633f5467383","Type":"ContainerDied","Data":"a1d699e97f4f3066e110d2189b65713edeea31f22ce56dbff4967e5ead216561"} Nov 28 11:37:50 crc kubenswrapper[4862]: I1128 11:37:50.082159 4862 generic.go:334] "Generic (PLEG): container finished" podID="95091706-4117-46dc-9405-f633f5467383" containerID="0c3895de8782968244bdf12dbbb2f46ea83c5d58791ac3126916ed1f9c02415d" exitCode=0 Nov 28 11:37:50 crc kubenswrapper[4862]: I1128 11:37:50.082234 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" event={"ID":"95091706-4117-46dc-9405-f633f5467383","Type":"ContainerDied","Data":"0c3895de8782968244bdf12dbbb2f46ea83c5d58791ac3126916ed1f9c02415d"} Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.481340 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.574972 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-util\") pod \"95091706-4117-46dc-9405-f633f5467383\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.575042 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-bundle\") pod \"95091706-4117-46dc-9405-f633f5467383\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.575084 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg8vc\" (UniqueName: \"kubernetes.io/projected/95091706-4117-46dc-9405-f633f5467383-kube-api-access-tg8vc\") pod \"95091706-4117-46dc-9405-f633f5467383\" (UID: \"95091706-4117-46dc-9405-f633f5467383\") " Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.576354 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-bundle" (OuterVolumeSpecName: "bundle") pod "95091706-4117-46dc-9405-f633f5467383" (UID: "95091706-4117-46dc-9405-f633f5467383"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.585371 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95091706-4117-46dc-9405-f633f5467383-kube-api-access-tg8vc" (OuterVolumeSpecName: "kube-api-access-tg8vc") pod "95091706-4117-46dc-9405-f633f5467383" (UID: "95091706-4117-46dc-9405-f633f5467383"). InnerVolumeSpecName "kube-api-access-tg8vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.588748 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-util" (OuterVolumeSpecName: "util") pod "95091706-4117-46dc-9405-f633f5467383" (UID: "95091706-4117-46dc-9405-f633f5467383"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.675986 4862 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-util\") on node \"crc\" DevicePath \"\"" Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.676055 4862 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95091706-4117-46dc-9405-f633f5467383-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:37:51 crc kubenswrapper[4862]: I1128 11:37:51.676073 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg8vc\" (UniqueName: \"kubernetes.io/projected/95091706-4117-46dc-9405-f633f5467383-kube-api-access-tg8vc\") on node \"crc\" DevicePath \"\"" Nov 28 11:37:52 crc kubenswrapper[4862]: I1128 11:37:52.096940 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" event={"ID":"95091706-4117-46dc-9405-f633f5467383","Type":"ContainerDied","Data":"f06be3bf5921e56b72b37038fc1bb6d85143561777d7b2706df8c8e7c6e51514"} Nov 28 11:37:52 crc kubenswrapper[4862]: I1128 11:37:52.097246 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f06be3bf5921e56b72b37038fc1bb6d85143561777d7b2706df8c8e7c6e51514" Nov 28 11:37:52 crc kubenswrapper[4862]: I1128 11:37:52.097027 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.301859 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g"] Nov 28 11:37:57 crc kubenswrapper[4862]: E1128 11:37:57.303441 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95091706-4117-46dc-9405-f633f5467383" containerName="extract" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.303528 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="95091706-4117-46dc-9405-f633f5467383" containerName="extract" Nov 28 11:37:57 crc kubenswrapper[4862]: E1128 11:37:57.303614 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95091706-4117-46dc-9405-f633f5467383" containerName="pull" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.303668 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="95091706-4117-46dc-9405-f633f5467383" containerName="pull" Nov 28 11:37:57 crc kubenswrapper[4862]: E1128 11:37:57.303732 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95091706-4117-46dc-9405-f633f5467383" containerName="util" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.303782 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="95091706-4117-46dc-9405-f633f5467383" containerName="util" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.303945 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="95091706-4117-46dc-9405-f633f5467383" containerName="extract" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.304418 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.306725 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.306958 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.307036 4862 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-9r7wd" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.313715 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g"] Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.359296 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6brnr\" (UniqueName: \"kubernetes.io/projected/873eb263-7ab2-4bdc-ae98-a58538256794-kube-api-access-6brnr\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwg6g\" (UID: \"873eb263-7ab2-4bdc-ae98-a58538256794\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.359431 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/873eb263-7ab2-4bdc-ae98-a58538256794-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwg6g\" (UID: \"873eb263-7ab2-4bdc-ae98-a58538256794\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.461043 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6brnr\" (UniqueName: \"kubernetes.io/projected/873eb263-7ab2-4bdc-ae98-a58538256794-kube-api-access-6brnr\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwg6g\" (UID: \"873eb263-7ab2-4bdc-ae98-a58538256794\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.461140 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/873eb263-7ab2-4bdc-ae98-a58538256794-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwg6g\" (UID: \"873eb263-7ab2-4bdc-ae98-a58538256794\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.461880 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/873eb263-7ab2-4bdc-ae98-a58538256794-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwg6g\" (UID: \"873eb263-7ab2-4bdc-ae98-a58538256794\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.481639 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6brnr\" (UniqueName: \"kubernetes.io/projected/873eb263-7ab2-4bdc-ae98-a58538256794-kube-api-access-6brnr\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwg6g\" (UID: \"873eb263-7ab2-4bdc-ae98-a58538256794\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" Nov 28 11:37:57 crc kubenswrapper[4862]: I1128 11:37:57.682957 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" Nov 28 11:37:58 crc kubenswrapper[4862]: I1128 11:37:58.219393 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g"] Nov 28 11:37:58 crc kubenswrapper[4862]: I1128 11:37:58.873927 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-tz4mw" Nov 28 11:37:59 crc kubenswrapper[4862]: I1128 11:37:59.140248 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" event={"ID":"873eb263-7ab2-4bdc-ae98-a58538256794","Type":"ContainerStarted","Data":"93173bab19426e2d99f5fc714a9e694a008d131bd127f9c407347fa244edf170"} Nov 28 11:38:15 crc kubenswrapper[4862]: E1128 11:38:15.031371 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:fa8de363ab4435c1085ac37f1bad488828c6ae8ba361c5f865c27ef577610911" Nov 28 11:38:15 crc kubenswrapper[4862]: E1128 11:38:15.031970 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-operator,Image:registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:fa8de363ab4435c1085ac37f1bad488828c6ae8ba361c5f865c27ef577610911,Command:[/usr/bin/cert-manager-operator],Args:[start --v=$(OPERATOR_LOG_LEVEL) --trusted-ca-configmap=$(TRUSTED_CA_CONFIGMAP_NAME) --cloud-credentials-secret=$(CLOUD_CREDENTIALS_SECRET_NAME) --unsupported-addon-features=$(UNSUPPORTED_ADDON_FEATURES)],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:cert-manager-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_WEBHOOK,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CA_INJECTOR,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CONTROLLER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ACMESOLVER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:ba937fc4b9eee31422914352c11a45b90754ba4fbe490ea45249b90afdc4e0a7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ISTIOCSR,Value:registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:af1ac813b8ee414ef215936f05197bc498bccbd540f3e2a93cb522221ba112bc,ValueFrom:nil,},EnvVar{Name:OPERAND_IMAGE_VERSION,Value:1.18.3,ValueFrom:nil,},EnvVar{Name:ISTIOCSR_OPERAND_IMAGE_VERSION,Value:0.14.2,ValueFrom:nil,},EnvVar{Name:OPERATOR_IMAGE_VERSION,Value:1.18.0,ValueFrom:nil,},EnvVar{Name:OPERATOR_LOG_LEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:TRUSTED_CA_CONFIGMAP_NAME,Value:,ValueFrom:nil,},EnvVar{Name:CLOUD_CREDENTIALS_SECRET_NAME,Value:,ValueFrom:nil,},EnvVar{Name:UNSUPPORTED_ADDON_FEATURES,Value:,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cert-manager-operator.v1.18.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{33554432 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tmp,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6brnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000700000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-operator-controller-manager-64cf6dff88-jwg6g_cert-manager-operator(873eb263-7ab2-4bdc-ae98-a58538256794): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 11:38:15 crc kubenswrapper[4862]: E1128 11:38:15.033423 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" podUID="873eb263-7ab2-4bdc-ae98-a58538256794" Nov 28 11:38:15 crc kubenswrapper[4862]: E1128 11:38:15.249823 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:fa8de363ab4435c1085ac37f1bad488828c6ae8ba361c5f865c27ef577610911\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" podUID="873eb263-7ab2-4bdc-ae98-a58538256794" Nov 28 11:38:28 crc kubenswrapper[4862]: I1128 11:38:28.348741 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" event={"ID":"873eb263-7ab2-4bdc-ae98-a58538256794","Type":"ContainerStarted","Data":"8b2dd2038e9acf328e9682059710c695a57b5b0a3512da24351b50fb5e6d4440"} Nov 28 11:38:28 crc kubenswrapper[4862]: I1128 11:38:28.390920 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwg6g" podStartSLOduration=2.07173659 podStartE2EDuration="31.390886043s" podCreationTimestamp="2025-11-28 11:37:57 +0000 UTC" firstStartedPulling="2025-11-28 11:37:58.234879278 +0000 UTC m=+955.767393199" lastFinishedPulling="2025-11-28 11:38:27.554028701 +0000 UTC m=+985.086542652" observedRunningTime="2025-11-28 11:38:28.389167204 +0000 UTC m=+985.921681145" watchObservedRunningTime="2025-11-28 11:38:28.390886043 +0000 UTC m=+985.923400004" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.628596 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-d9zx4"] Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.629858 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.631401 4862 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-4prcd" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.632619 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.634407 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.670084 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wnp9\" (UniqueName: \"kubernetes.io/projected/a59357d2-2d8f-4701-a716-2253d0f35786-kube-api-access-9wnp9\") pod \"cert-manager-webhook-f4fb5df64-d9zx4\" (UID: \"a59357d2-2d8f-4701-a716-2253d0f35786\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.670206 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a59357d2-2d8f-4701-a716-2253d0f35786-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-d9zx4\" (UID: \"a59357d2-2d8f-4701-a716-2253d0f35786\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.693793 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-d9zx4"] Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.771441 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wnp9\" (UniqueName: \"kubernetes.io/projected/a59357d2-2d8f-4701-a716-2253d0f35786-kube-api-access-9wnp9\") pod \"cert-manager-webhook-f4fb5df64-d9zx4\" (UID: \"a59357d2-2d8f-4701-a716-2253d0f35786\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.771503 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a59357d2-2d8f-4701-a716-2253d0f35786-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-d9zx4\" (UID: \"a59357d2-2d8f-4701-a716-2253d0f35786\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.794573 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wnp9\" (UniqueName: \"kubernetes.io/projected/a59357d2-2d8f-4701-a716-2253d0f35786-kube-api-access-9wnp9\") pod \"cert-manager-webhook-f4fb5df64-d9zx4\" (UID: \"a59357d2-2d8f-4701-a716-2253d0f35786\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.798498 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a59357d2-2d8f-4701-a716-2253d0f35786-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-d9zx4\" (UID: \"a59357d2-2d8f-4701-a716-2253d0f35786\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:31 crc kubenswrapper[4862]: I1128 11:38:31.949871 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.266046 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-d9zx4"] Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.372410 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" event={"ID":"a59357d2-2d8f-4701-a716-2253d0f35786","Type":"ContainerStarted","Data":"0a3a02c8d36326d9dc3036cd8b3ebe45c51a7e36face9e41c42adc2441dc4e03"} Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.610676 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5"] Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.611379 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.613680 4862 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-t2bls" Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.647221 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5"] Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.684006 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d559bcd-98f2-43ea-9f49-e2a2a8d39c32-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-xrhk5\" (UID: \"5d559bcd-98f2-43ea-9f49-e2a2a8d39c32\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.684083 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwknf\" (UniqueName: \"kubernetes.io/projected/5d559bcd-98f2-43ea-9f49-e2a2a8d39c32-kube-api-access-qwknf\") pod \"cert-manager-cainjector-855d9ccff4-xrhk5\" (UID: \"5d559bcd-98f2-43ea-9f49-e2a2a8d39c32\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.785902 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d559bcd-98f2-43ea-9f49-e2a2a8d39c32-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-xrhk5\" (UID: \"5d559bcd-98f2-43ea-9f49-e2a2a8d39c32\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.786345 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwknf\" (UniqueName: \"kubernetes.io/projected/5d559bcd-98f2-43ea-9f49-e2a2a8d39c32-kube-api-access-qwknf\") pod \"cert-manager-cainjector-855d9ccff4-xrhk5\" (UID: \"5d559bcd-98f2-43ea-9f49-e2a2a8d39c32\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.816109 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwknf\" (UniqueName: \"kubernetes.io/projected/5d559bcd-98f2-43ea-9f49-e2a2a8d39c32-kube-api-access-qwknf\") pod \"cert-manager-cainjector-855d9ccff4-xrhk5\" (UID: \"5d559bcd-98f2-43ea-9f49-e2a2a8d39c32\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.837418 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d559bcd-98f2-43ea-9f49-e2a2a8d39c32-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-xrhk5\" (UID: \"5d559bcd-98f2-43ea-9f49-e2a2a8d39c32\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" Nov 28 11:38:32 crc kubenswrapper[4862]: I1128 11:38:32.942221 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" Nov 28 11:38:33 crc kubenswrapper[4862]: I1128 11:38:33.202879 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5"] Nov 28 11:38:33 crc kubenswrapper[4862]: W1128 11:38:33.233710 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d559bcd_98f2_43ea_9f49_e2a2a8d39c32.slice/crio-aa9533eb45d6747df6cfa7b1c77e389cd1dea112a7c02b188cdf48b833218961 WatchSource:0}: Error finding container aa9533eb45d6747df6cfa7b1c77e389cd1dea112a7c02b188cdf48b833218961: Status 404 returned error can't find the container with id aa9533eb45d6747df6cfa7b1c77e389cd1dea112a7c02b188cdf48b833218961 Nov 28 11:38:33 crc kubenswrapper[4862]: I1128 11:38:33.381155 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" event={"ID":"5d559bcd-98f2-43ea-9f49-e2a2a8d39c32","Type":"ContainerStarted","Data":"aa9533eb45d6747df6cfa7b1c77e389cd1dea112a7c02b188cdf48b833218961"} Nov 28 11:38:38 crc kubenswrapper[4862]: I1128 11:38:38.292574 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:38:38 crc kubenswrapper[4862]: I1128 11:38:38.293572 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:38:42 crc kubenswrapper[4862]: I1128 11:38:42.462233 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" event={"ID":"5d559bcd-98f2-43ea-9f49-e2a2a8d39c32","Type":"ContainerStarted","Data":"ecbb1ec53e5b4144795efc719e76e345cb02433af4fdb510e5309cf3319060ca"} Nov 28 11:38:42 crc kubenswrapper[4862]: I1128 11:38:42.463569 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" event={"ID":"a59357d2-2d8f-4701-a716-2253d0f35786","Type":"ContainerStarted","Data":"1bc1c54e4b525f3bdbde1b7dfdc23bae3215346eac5537b1542640ca786a65e9"} Nov 28 11:38:42 crc kubenswrapper[4862]: I1128 11:38:42.463744 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:42 crc kubenswrapper[4862]: I1128 11:38:42.499799 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-xrhk5" podStartSLOduration=2.00645897 podStartE2EDuration="10.49977771s" podCreationTimestamp="2025-11-28 11:38:32 +0000 UTC" firstStartedPulling="2025-11-28 11:38:33.244313821 +0000 UTC m=+990.776827742" lastFinishedPulling="2025-11-28 11:38:41.737632551 +0000 UTC m=+999.270146482" observedRunningTime="2025-11-28 11:38:42.490600496 +0000 UTC m=+1000.023114417" watchObservedRunningTime="2025-11-28 11:38:42.49977771 +0000 UTC m=+1000.032291631" Nov 28 11:38:42 crc kubenswrapper[4862]: I1128 11:38:42.511071 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" podStartSLOduration=2.043360452 podStartE2EDuration="11.511043855s" podCreationTimestamp="2025-11-28 11:38:31 +0000 UTC" firstStartedPulling="2025-11-28 11:38:32.278946577 +0000 UTC m=+989.811460498" lastFinishedPulling="2025-11-28 11:38:41.74662998 +0000 UTC m=+999.279143901" observedRunningTime="2025-11-28 11:38:42.505044562 +0000 UTC m=+1000.037558483" watchObservedRunningTime="2025-11-28 11:38:42.511043855 +0000 UTC m=+1000.043557776" Nov 28 11:38:46 crc kubenswrapper[4862]: I1128 11:38:46.953925 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-d9zx4" Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.573322 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-cml9m"] Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.574555 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-cml9m" Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.580061 4862 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-lpt9k" Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.592739 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-cml9m"] Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.640478 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c482\" (UniqueName: \"kubernetes.io/projected/43876f80-dd71-47f5-8b77-f22fb71e23f4-kube-api-access-2c482\") pod \"cert-manager-86cb77c54b-cml9m\" (UID: \"43876f80-dd71-47f5-8b77-f22fb71e23f4\") " pod="cert-manager/cert-manager-86cb77c54b-cml9m" Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.640663 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/43876f80-dd71-47f5-8b77-f22fb71e23f4-bound-sa-token\") pod \"cert-manager-86cb77c54b-cml9m\" (UID: \"43876f80-dd71-47f5-8b77-f22fb71e23f4\") " pod="cert-manager/cert-manager-86cb77c54b-cml9m" Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.742903 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c482\" (UniqueName: \"kubernetes.io/projected/43876f80-dd71-47f5-8b77-f22fb71e23f4-kube-api-access-2c482\") pod \"cert-manager-86cb77c54b-cml9m\" (UID: \"43876f80-dd71-47f5-8b77-f22fb71e23f4\") " pod="cert-manager/cert-manager-86cb77c54b-cml9m" Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.743086 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/43876f80-dd71-47f5-8b77-f22fb71e23f4-bound-sa-token\") pod \"cert-manager-86cb77c54b-cml9m\" (UID: \"43876f80-dd71-47f5-8b77-f22fb71e23f4\") " pod="cert-manager/cert-manager-86cb77c54b-cml9m" Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.770043 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/43876f80-dd71-47f5-8b77-f22fb71e23f4-bound-sa-token\") pod \"cert-manager-86cb77c54b-cml9m\" (UID: \"43876f80-dd71-47f5-8b77-f22fb71e23f4\") " pod="cert-manager/cert-manager-86cb77c54b-cml9m" Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.772357 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c482\" (UniqueName: \"kubernetes.io/projected/43876f80-dd71-47f5-8b77-f22fb71e23f4-kube-api-access-2c482\") pod \"cert-manager-86cb77c54b-cml9m\" (UID: \"43876f80-dd71-47f5-8b77-f22fb71e23f4\") " pod="cert-manager/cert-manager-86cb77c54b-cml9m" Nov 28 11:38:49 crc kubenswrapper[4862]: I1128 11:38:49.903317 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-cml9m" Nov 28 11:38:50 crc kubenswrapper[4862]: I1128 11:38:50.345214 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-cml9m"] Nov 28 11:38:50 crc kubenswrapper[4862]: I1128 11:38:50.519472 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-cml9m" event={"ID":"43876f80-dd71-47f5-8b77-f22fb71e23f4","Type":"ContainerStarted","Data":"dfbcd4adc8bfe8d901b2205f39f3d91868d5140face9bf3b85ff51acf427c2ec"} Nov 28 11:38:52 crc kubenswrapper[4862]: I1128 11:38:52.534442 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-cml9m" event={"ID":"43876f80-dd71-47f5-8b77-f22fb71e23f4","Type":"ContainerStarted","Data":"71094b8c38efb9219f9449c3ee04af8aaf3197916dc2e2434bc4eb3088cbbfd5"} Nov 28 11:38:52 crc kubenswrapper[4862]: I1128 11:38:52.558620 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-cml9m" podStartSLOduration=3.558600803 podStartE2EDuration="3.558600803s" podCreationTimestamp="2025-11-28 11:38:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:38:52.555010179 +0000 UTC m=+1010.087524100" watchObservedRunningTime="2025-11-28 11:38:52.558600803 +0000 UTC m=+1010.091114724" Nov 28 11:39:00 crc kubenswrapper[4862]: I1128 11:39:00.911759 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2gqx2"] Nov 28 11:39:00 crc kubenswrapper[4862]: I1128 11:39:00.914971 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2gqx2" Nov 28 11:39:00 crc kubenswrapper[4862]: I1128 11:39:00.917190 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 28 11:39:00 crc kubenswrapper[4862]: I1128 11:39:00.917891 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-dmtnw" Nov 28 11:39:00 crc kubenswrapper[4862]: I1128 11:39:00.921690 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 28 11:39:00 crc kubenswrapper[4862]: I1128 11:39:00.926688 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2gqx2"] Nov 28 11:39:00 crc kubenswrapper[4862]: I1128 11:39:00.971226 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt5sf\" (UniqueName: \"kubernetes.io/projected/e244afc0-2b04-42fd-b149-2fb1d0739f71-kube-api-access-gt5sf\") pod \"openstack-operator-index-2gqx2\" (UID: \"e244afc0-2b04-42fd-b149-2fb1d0739f71\") " pod="openstack-operators/openstack-operator-index-2gqx2" Nov 28 11:39:01 crc kubenswrapper[4862]: I1128 11:39:01.072251 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt5sf\" (UniqueName: \"kubernetes.io/projected/e244afc0-2b04-42fd-b149-2fb1d0739f71-kube-api-access-gt5sf\") pod \"openstack-operator-index-2gqx2\" (UID: \"e244afc0-2b04-42fd-b149-2fb1d0739f71\") " pod="openstack-operators/openstack-operator-index-2gqx2" Nov 28 11:39:01 crc kubenswrapper[4862]: I1128 11:39:01.096314 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt5sf\" (UniqueName: \"kubernetes.io/projected/e244afc0-2b04-42fd-b149-2fb1d0739f71-kube-api-access-gt5sf\") pod \"openstack-operator-index-2gqx2\" (UID: \"e244afc0-2b04-42fd-b149-2fb1d0739f71\") " pod="openstack-operators/openstack-operator-index-2gqx2" Nov 28 11:39:01 crc kubenswrapper[4862]: I1128 11:39:01.249991 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2gqx2" Nov 28 11:39:01 crc kubenswrapper[4862]: I1128 11:39:01.653931 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2gqx2"] Nov 28 11:39:02 crc kubenswrapper[4862]: I1128 11:39:02.603669 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2gqx2" event={"ID":"e244afc0-2b04-42fd-b149-2fb1d0739f71","Type":"ContainerStarted","Data":"94bb70d1cc493fe2c0eeeab1d867d0825393a599fb42f668d7615ce64c367b83"} Nov 28 11:39:03 crc kubenswrapper[4862]: I1128 11:39:03.612481 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2gqx2" event={"ID":"e244afc0-2b04-42fd-b149-2fb1d0739f71","Type":"ContainerStarted","Data":"4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a"} Nov 28 11:39:03 crc kubenswrapper[4862]: I1128 11:39:03.630307 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2gqx2" podStartSLOduration=2.5169935839999997 podStartE2EDuration="3.630285841s" podCreationTimestamp="2025-11-28 11:39:00 +0000 UTC" firstStartedPulling="2025-11-28 11:39:01.667749746 +0000 UTC m=+1019.200263667" lastFinishedPulling="2025-11-28 11:39:02.781041963 +0000 UTC m=+1020.313555924" observedRunningTime="2025-11-28 11:39:03.629782726 +0000 UTC m=+1021.162296687" watchObservedRunningTime="2025-11-28 11:39:03.630285841 +0000 UTC m=+1021.162799782" Nov 28 11:39:04 crc kubenswrapper[4862]: I1128 11:39:04.091268 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2gqx2"] Nov 28 11:39:04 crc kubenswrapper[4862]: I1128 11:39:04.702037 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-g64fp"] Nov 28 11:39:04 crc kubenswrapper[4862]: I1128 11:39:04.704301 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g64fp" Nov 28 11:39:04 crc kubenswrapper[4862]: I1128 11:39:04.719631 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g64fp"] Nov 28 11:39:04 crc kubenswrapper[4862]: I1128 11:39:04.725447 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnc8w\" (UniqueName: \"kubernetes.io/projected/e37fdd75-4a2c-4758-a2de-9dd365f7ead6-kube-api-access-hnc8w\") pod \"openstack-operator-index-g64fp\" (UID: \"e37fdd75-4a2c-4758-a2de-9dd365f7ead6\") " pod="openstack-operators/openstack-operator-index-g64fp" Nov 28 11:39:04 crc kubenswrapper[4862]: I1128 11:39:04.826839 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnc8w\" (UniqueName: \"kubernetes.io/projected/e37fdd75-4a2c-4758-a2de-9dd365f7ead6-kube-api-access-hnc8w\") pod \"openstack-operator-index-g64fp\" (UID: \"e37fdd75-4a2c-4758-a2de-9dd365f7ead6\") " pod="openstack-operators/openstack-operator-index-g64fp" Nov 28 11:39:04 crc kubenswrapper[4862]: I1128 11:39:04.852450 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnc8w\" (UniqueName: \"kubernetes.io/projected/e37fdd75-4a2c-4758-a2de-9dd365f7ead6-kube-api-access-hnc8w\") pod \"openstack-operator-index-g64fp\" (UID: \"e37fdd75-4a2c-4758-a2de-9dd365f7ead6\") " pod="openstack-operators/openstack-operator-index-g64fp" Nov 28 11:39:05 crc kubenswrapper[4862]: I1128 11:39:05.047807 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g64fp" Nov 28 11:39:05 crc kubenswrapper[4862]: I1128 11:39:05.560384 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g64fp"] Nov 28 11:39:05 crc kubenswrapper[4862]: I1128 11:39:05.627165 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2gqx2" podUID="e244afc0-2b04-42fd-b149-2fb1d0739f71" containerName="registry-server" containerID="cri-o://4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a" gracePeriod=2 Nov 28 11:39:05 crc kubenswrapper[4862]: I1128 11:39:05.627378 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g64fp" event={"ID":"e37fdd75-4a2c-4758-a2de-9dd365f7ead6","Type":"ContainerStarted","Data":"bc9993ff91e984b20fbb5db3440f6a36344482ea9af4734e7b66cdf6d32a15ef"} Nov 28 11:39:05 crc kubenswrapper[4862]: I1128 11:39:05.989952 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2gqx2" Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.147236 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt5sf\" (UniqueName: \"kubernetes.io/projected/e244afc0-2b04-42fd-b149-2fb1d0739f71-kube-api-access-gt5sf\") pod \"e244afc0-2b04-42fd-b149-2fb1d0739f71\" (UID: \"e244afc0-2b04-42fd-b149-2fb1d0739f71\") " Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.156512 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e244afc0-2b04-42fd-b149-2fb1d0739f71-kube-api-access-gt5sf" (OuterVolumeSpecName: "kube-api-access-gt5sf") pod "e244afc0-2b04-42fd-b149-2fb1d0739f71" (UID: "e244afc0-2b04-42fd-b149-2fb1d0739f71"). InnerVolumeSpecName "kube-api-access-gt5sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.249369 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt5sf\" (UniqueName: \"kubernetes.io/projected/e244afc0-2b04-42fd-b149-2fb1d0739f71-kube-api-access-gt5sf\") on node \"crc\" DevicePath \"\"" Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.637228 4862 generic.go:334] "Generic (PLEG): container finished" podID="e244afc0-2b04-42fd-b149-2fb1d0739f71" containerID="4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a" exitCode=0 Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.637308 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2gqx2" event={"ID":"e244afc0-2b04-42fd-b149-2fb1d0739f71","Type":"ContainerDied","Data":"4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a"} Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.637354 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2gqx2" Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.637373 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2gqx2" event={"ID":"e244afc0-2b04-42fd-b149-2fb1d0739f71","Type":"ContainerDied","Data":"94bb70d1cc493fe2c0eeeab1d867d0825393a599fb42f668d7615ce64c367b83"} Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.637391 4862 scope.go:117] "RemoveContainer" containerID="4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a" Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.639961 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g64fp" event={"ID":"e37fdd75-4a2c-4758-a2de-9dd365f7ead6","Type":"ContainerStarted","Data":"108bbea4393c1f7f7789bd7d4bbe6507fc5866c4e447cb869beb2966b2bbadec"} Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.661118 4862 scope.go:117] "RemoveContainer" containerID="4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a" Nov 28 11:39:06 crc kubenswrapper[4862]: E1128 11:39:06.662697 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a\": container with ID starting with 4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a not found: ID does not exist" containerID="4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a" Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.662731 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a"} err="failed to get container status \"4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a\": rpc error: code = NotFound desc = could not find container \"4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a\": container with ID starting with 4bb68f70cbba2a16034d61851b3112faf81335e2591db046d45891c7f0f9925a not found: ID does not exist" Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.663949 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-g64fp" podStartSLOduration=2.188358915 podStartE2EDuration="2.663938697s" podCreationTimestamp="2025-11-28 11:39:04 +0000 UTC" firstStartedPulling="2025-11-28 11:39:05.570255256 +0000 UTC m=+1023.102769177" lastFinishedPulling="2025-11-28 11:39:06.045835038 +0000 UTC m=+1023.578348959" observedRunningTime="2025-11-28 11:39:06.663879975 +0000 UTC m=+1024.196393896" watchObservedRunningTime="2025-11-28 11:39:06.663938697 +0000 UTC m=+1024.196452618" Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.702284 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2gqx2"] Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.711622 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-2gqx2"] Nov 28 11:39:06 crc kubenswrapper[4862]: I1128 11:39:06.845891 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e244afc0-2b04-42fd-b149-2fb1d0739f71" path="/var/lib/kubelet/pods/e244afc0-2b04-42fd-b149-2fb1d0739f71/volumes" Nov 28 11:39:08 crc kubenswrapper[4862]: I1128 11:39:08.292170 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:39:08 crc kubenswrapper[4862]: I1128 11:39:08.292719 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:39:15 crc kubenswrapper[4862]: I1128 11:39:15.048778 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-g64fp" Nov 28 11:39:15 crc kubenswrapper[4862]: I1128 11:39:15.050236 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-g64fp" Nov 28 11:39:15 crc kubenswrapper[4862]: I1128 11:39:15.090412 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-g64fp" Nov 28 11:39:15 crc kubenswrapper[4862]: I1128 11:39:15.749698 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-g64fp" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.337354 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6"] Nov 28 11:39:22 crc kubenswrapper[4862]: E1128 11:39:22.339033 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e244afc0-2b04-42fd-b149-2fb1d0739f71" containerName="registry-server" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.339125 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e244afc0-2b04-42fd-b149-2fb1d0739f71" containerName="registry-server" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.339324 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e244afc0-2b04-42fd-b149-2fb1d0739f71" containerName="registry-server" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.340339 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.342651 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-gfszb" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.351229 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6"] Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.407553 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-bundle\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.407616 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbn6d\" (UniqueName: \"kubernetes.io/projected/d23c71bc-2135-4017-a1f7-dabbf039770b-kube-api-access-tbn6d\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.407643 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-util\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.508410 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-bundle\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.508477 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbn6d\" (UniqueName: \"kubernetes.io/projected/d23c71bc-2135-4017-a1f7-dabbf039770b-kube-api-access-tbn6d\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.508501 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-util\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.509055 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-util\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.509116 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-bundle\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.528217 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbn6d\" (UniqueName: \"kubernetes.io/projected/d23c71bc-2135-4017-a1f7-dabbf039770b-kube-api-access-tbn6d\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:22 crc kubenswrapper[4862]: I1128 11:39:22.657853 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:23 crc kubenswrapper[4862]: I1128 11:39:23.078642 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6"] Nov 28 11:39:23 crc kubenswrapper[4862]: I1128 11:39:23.780304 4862 generic.go:334] "Generic (PLEG): container finished" podID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerID="05858325c18d3bd3cd3f48cb2c31583b1f56602356d9f64b819202f466a69ceb" exitCode=0 Nov 28 11:39:23 crc kubenswrapper[4862]: I1128 11:39:23.780395 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" event={"ID":"d23c71bc-2135-4017-a1f7-dabbf039770b","Type":"ContainerDied","Data":"05858325c18d3bd3cd3f48cb2c31583b1f56602356d9f64b819202f466a69ceb"} Nov 28 11:39:23 crc kubenswrapper[4862]: I1128 11:39:23.780636 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" event={"ID":"d23c71bc-2135-4017-a1f7-dabbf039770b","Type":"ContainerStarted","Data":"dd37e4cea2b02aa02d7275e6b50a904fa0fc3b183b137131b3ce590e2702cf9b"} Nov 28 11:39:24 crc kubenswrapper[4862]: I1128 11:39:24.791167 4862 generic.go:334] "Generic (PLEG): container finished" podID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerID="a5ba6de9ba8a262dd9f75e2ac110fa1b8e26edfb8c887432dd12742f9fc6ee88" exitCode=0 Nov 28 11:39:24 crc kubenswrapper[4862]: I1128 11:39:24.791308 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" event={"ID":"d23c71bc-2135-4017-a1f7-dabbf039770b","Type":"ContainerDied","Data":"a5ba6de9ba8a262dd9f75e2ac110fa1b8e26edfb8c887432dd12742f9fc6ee88"} Nov 28 11:39:25 crc kubenswrapper[4862]: I1128 11:39:25.812291 4862 generic.go:334] "Generic (PLEG): container finished" podID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerID="b77e8e0911cc50fa4925a92859ab3a4ff23ea961564443f9b038723a05cd8915" exitCode=0 Nov 28 11:39:25 crc kubenswrapper[4862]: I1128 11:39:25.812340 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" event={"ID":"d23c71bc-2135-4017-a1f7-dabbf039770b","Type":"ContainerDied","Data":"b77e8e0911cc50fa4925a92859ab3a4ff23ea961564443f9b038723a05cd8915"} Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.076210 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.174493 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbn6d\" (UniqueName: \"kubernetes.io/projected/d23c71bc-2135-4017-a1f7-dabbf039770b-kube-api-access-tbn6d\") pod \"d23c71bc-2135-4017-a1f7-dabbf039770b\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.174531 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-bundle\") pod \"d23c71bc-2135-4017-a1f7-dabbf039770b\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.174581 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-util\") pod \"d23c71bc-2135-4017-a1f7-dabbf039770b\" (UID: \"d23c71bc-2135-4017-a1f7-dabbf039770b\") " Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.175500 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-bundle" (OuterVolumeSpecName: "bundle") pod "d23c71bc-2135-4017-a1f7-dabbf039770b" (UID: "d23c71bc-2135-4017-a1f7-dabbf039770b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.186343 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d23c71bc-2135-4017-a1f7-dabbf039770b-kube-api-access-tbn6d" (OuterVolumeSpecName: "kube-api-access-tbn6d") pod "d23c71bc-2135-4017-a1f7-dabbf039770b" (UID: "d23c71bc-2135-4017-a1f7-dabbf039770b"). InnerVolumeSpecName "kube-api-access-tbn6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.191846 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-util" (OuterVolumeSpecName: "util") pod "d23c71bc-2135-4017-a1f7-dabbf039770b" (UID: "d23c71bc-2135-4017-a1f7-dabbf039770b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.276026 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbn6d\" (UniqueName: \"kubernetes.io/projected/d23c71bc-2135-4017-a1f7-dabbf039770b-kube-api-access-tbn6d\") on node \"crc\" DevicePath \"\"" Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.276115 4862 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.276143 4862 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d23c71bc-2135-4017-a1f7-dabbf039770b-util\") on node \"crc\" DevicePath \"\"" Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.831041 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" event={"ID":"d23c71bc-2135-4017-a1f7-dabbf039770b","Type":"ContainerDied","Data":"dd37e4cea2b02aa02d7275e6b50a904fa0fc3b183b137131b3ce590e2702cf9b"} Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.831289 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd37e4cea2b02aa02d7275e6b50a904fa0fc3b183b137131b3ce590e2702cf9b" Nov 28 11:39:27 crc kubenswrapper[4862]: I1128 11:39:27.831399 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.543342 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q"] Nov 28 11:39:29 crc kubenswrapper[4862]: E1128 11:39:29.544034 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerName="util" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.544054 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerName="util" Nov 28 11:39:29 crc kubenswrapper[4862]: E1128 11:39:29.544120 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerName="extract" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.544133 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerName="extract" Nov 28 11:39:29 crc kubenswrapper[4862]: E1128 11:39:29.544156 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerName="pull" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.544168 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerName="pull" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.544361 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="d23c71bc-2135-4017-a1f7-dabbf039770b" containerName="extract" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.544990 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.547283 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-5h6z8" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.563054 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q"] Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.605523 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm48c\" (UniqueName: \"kubernetes.io/projected/dd995aaf-688a-4234-8132-b2ea19bd9fff-kube-api-access-zm48c\") pod \"openstack-operator-controller-operator-848c47db98-xgf7q\" (UID: \"dd995aaf-688a-4234-8132-b2ea19bd9fff\") " pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.706618 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm48c\" (UniqueName: \"kubernetes.io/projected/dd995aaf-688a-4234-8132-b2ea19bd9fff-kube-api-access-zm48c\") pod \"openstack-operator-controller-operator-848c47db98-xgf7q\" (UID: \"dd995aaf-688a-4234-8132-b2ea19bd9fff\") " pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.723613 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm48c\" (UniqueName: \"kubernetes.io/projected/dd995aaf-688a-4234-8132-b2ea19bd9fff-kube-api-access-zm48c\") pod \"openstack-operator-controller-operator-848c47db98-xgf7q\" (UID: \"dd995aaf-688a-4234-8132-b2ea19bd9fff\") " pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" Nov 28 11:39:29 crc kubenswrapper[4862]: I1128 11:39:29.874899 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" Nov 28 11:39:30 crc kubenswrapper[4862]: I1128 11:39:30.333782 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q"] Nov 28 11:39:30 crc kubenswrapper[4862]: I1128 11:39:30.850318 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" event={"ID":"dd995aaf-688a-4234-8132-b2ea19bd9fff","Type":"ContainerStarted","Data":"065b33af1d0457b74a9983e3cadfbd0b71cd347551fff3b95813c5467a733893"} Nov 28 11:39:35 crc kubenswrapper[4862]: I1128 11:39:35.889373 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" event={"ID":"dd995aaf-688a-4234-8132-b2ea19bd9fff","Type":"ContainerStarted","Data":"49563d7fe7e06d4e2936dc9e9704a16bb23ddb1ae5217c7bc2696d722dcb96dd"} Nov 28 11:39:35 crc kubenswrapper[4862]: I1128 11:39:35.889976 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" Nov 28 11:39:35 crc kubenswrapper[4862]: I1128 11:39:35.923512 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" podStartSLOduration=1.996189288 podStartE2EDuration="6.923479893s" podCreationTimestamp="2025-11-28 11:39:29 +0000 UTC" firstStartedPulling="2025-11-28 11:39:30.34095639 +0000 UTC m=+1047.873470311" lastFinishedPulling="2025-11-28 11:39:35.268246995 +0000 UTC m=+1052.800760916" observedRunningTime="2025-11-28 11:39:35.919489359 +0000 UTC m=+1053.452003300" watchObservedRunningTime="2025-11-28 11:39:35.923479893 +0000 UTC m=+1053.455993854" Nov 28 11:39:38 crc kubenswrapper[4862]: I1128 11:39:38.291711 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:39:38 crc kubenswrapper[4862]: I1128 11:39:38.292270 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:39:38 crc kubenswrapper[4862]: I1128 11:39:38.292356 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:39:38 crc kubenswrapper[4862]: I1128 11:39:38.293467 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e63a1b14a90646d7f5b8ec0b596a3abf9ade9420df0f277e56e83cc3a51ca76"} pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:39:38 crc kubenswrapper[4862]: I1128 11:39:38.293575 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" containerID="cri-o://5e63a1b14a90646d7f5b8ec0b596a3abf9ade9420df0f277e56e83cc3a51ca76" gracePeriod=600 Nov 28 11:39:38 crc kubenswrapper[4862]: I1128 11:39:38.915035 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6845439-1a4c-4f28-9ba4-29149d545995" containerID="5e63a1b14a90646d7f5b8ec0b596a3abf9ade9420df0f277e56e83cc3a51ca76" exitCode=0 Nov 28 11:39:38 crc kubenswrapper[4862]: I1128 11:39:38.915402 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerDied","Data":"5e63a1b14a90646d7f5b8ec0b596a3abf9ade9420df0f277e56e83cc3a51ca76"} Nov 28 11:39:38 crc kubenswrapper[4862]: I1128 11:39:38.915432 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"1439dec9a32df00b0169c77b7376186b3da92973a1be5f562a161099f8fb96c8"} Nov 28 11:39:38 crc kubenswrapper[4862]: I1128 11:39:38.915454 4862 scope.go:117] "RemoveContainer" containerID="cd460d7f02e8c5ea1bc5f7ee547d7f0b295f608f496d14e974dc72d5aea4821d" Nov 28 11:39:49 crc kubenswrapper[4862]: I1128 11:39:49.879337 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-xgf7q" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.583603 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.584999 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.589400 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.590342 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.593810 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-ll6bl" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.593826 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-zw2fc" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.602546 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.608022 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.613893 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-5qbwv"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.615075 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.624814 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dj7dh" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.647615 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-5qbwv"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.666451 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.667629 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.670487 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-lwx5d" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.671290 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.672868 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.679498 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.681431 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.689418 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.692410 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-vpj7g" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.692787 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-c6n5t" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.707958 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.743317 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7rnl\" (UniqueName: \"kubernetes.io/projected/8bba217a-0d90-4544-b772-62e6e5e58d74-kube-api-access-x7rnl\") pod \"barbican-operator-controller-manager-7b64f4fb85-dj8zz\" (UID: \"8bba217a-0d90-4544-b772-62e6e5e58d74\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.743392 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djmjc\" (UniqueName: \"kubernetes.io/projected/b1165cf7-ed1f-48ba-a6dc-1cab53edf166-kube-api-access-djmjc\") pod \"designate-operator-controller-manager-955677c94-5qbwv\" (UID: \"b1165cf7-ed1f-48ba-a6dc-1cab53edf166\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.743422 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjgvr\" (UniqueName: \"kubernetes.io/projected/f8ce3d3e-999e-48bf-887a-add2b478966d-kube-api-access-pjgvr\") pod \"cinder-operator-controller-manager-6b7f75547b-7vhdh\" (UID: \"f8ce3d3e-999e-48bf-887a-add2b478966d\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.778806 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.804435 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.805889 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.814281 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-9cvjt" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.814516 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.815717 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.830833 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.832221 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.834468 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-858x5" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.845565 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.846748 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdtvr\" (UniqueName: \"kubernetes.io/projected/50269334-ed51-4bfc-8b87-1243d75774a3-kube-api-access-fdtvr\") pod \"heat-operator-controller-manager-5b77f656f-v5t45\" (UID: \"50269334-ed51-4bfc-8b87-1243d75774a3\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.846859 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7rnl\" (UniqueName: \"kubernetes.io/projected/8bba217a-0d90-4544-b772-62e6e5e58d74-kube-api-access-x7rnl\") pod \"barbican-operator-controller-manager-7b64f4fb85-dj8zz\" (UID: \"8bba217a-0d90-4544-b772-62e6e5e58d74\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.846897 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djmjc\" (UniqueName: \"kubernetes.io/projected/b1165cf7-ed1f-48ba-a6dc-1cab53edf166-kube-api-access-djmjc\") pod \"designate-operator-controller-manager-955677c94-5qbwv\" (UID: \"b1165cf7-ed1f-48ba-a6dc-1cab53edf166\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.846922 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjgvr\" (UniqueName: \"kubernetes.io/projected/f8ce3d3e-999e-48bf-887a-add2b478966d-kube-api-access-pjgvr\") pod \"cinder-operator-controller-manager-6b7f75547b-7vhdh\" (UID: \"f8ce3d3e-999e-48bf-887a-add2b478966d\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.846957 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csrhw\" (UniqueName: \"kubernetes.io/projected/f33dc056-b642-4cf9-95d4-9c40ac0468ca-kube-api-access-csrhw\") pod \"horizon-operator-controller-manager-5d494799bf-lnvbp\" (UID: \"f33dc056-b642-4cf9-95d4-9c40ac0468ca\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.846984 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xlxd\" (UniqueName: \"kubernetes.io/projected/04fbac3c-ab55-4790-85b2-3269762d6a55-kube-api-access-6xlxd\") pod \"glance-operator-controller-manager-589cbd6b5b-6h5f6\" (UID: \"04fbac3c-ab55-4790-85b2-3269762d6a55\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.857215 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.858302 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.866779 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-ccwfl" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.878161 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.886782 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djmjc\" (UniqueName: \"kubernetes.io/projected/b1165cf7-ed1f-48ba-a6dc-1cab53edf166-kube-api-access-djmjc\") pod \"designate-operator-controller-manager-955677c94-5qbwv\" (UID: \"b1165cf7-ed1f-48ba-a6dc-1cab53edf166\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.889029 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7rnl\" (UniqueName: \"kubernetes.io/projected/8bba217a-0d90-4544-b772-62e6e5e58d74-kube-api-access-x7rnl\") pod \"barbican-operator-controller-manager-7b64f4fb85-dj8zz\" (UID: \"8bba217a-0d90-4544-b772-62e6e5e58d74\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.889876 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.890894 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.893159 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjgvr\" (UniqueName: \"kubernetes.io/projected/f8ce3d3e-999e-48bf-887a-add2b478966d-kube-api-access-pjgvr\") pod \"cinder-operator-controller-manager-6b7f75547b-7vhdh\" (UID: \"f8ce3d3e-999e-48bf-887a-add2b478966d\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.893322 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-kcvqq" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.907407 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.919542 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.934822 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.935919 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.936645 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.937932 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-r2rbb" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.947829 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4"] Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.948892 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.949887 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhfwj\" (UniqueName: \"kubernetes.io/projected/c50b92c1-e947-4f41-abfe-278e1826a0c9-kube-api-access-jhfwj\") pod \"ironic-operator-controller-manager-67cb4dc6d4-dmqbn\" (UID: \"c50b92c1-e947-4f41-abfe-278e1826a0c9\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.949917 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdtvr\" (UniqueName: \"kubernetes.io/projected/50269334-ed51-4bfc-8b87-1243d75774a3-kube-api-access-fdtvr\") pod \"heat-operator-controller-manager-5b77f656f-v5t45\" (UID: \"50269334-ed51-4bfc-8b87-1243d75774a3\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.949960 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2lw2\" (UniqueName: \"kubernetes.io/projected/f4857ca4-a829-4205-b776-fa73faf12eac-kube-api-access-c2lw2\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.950003 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csrhw\" (UniqueName: \"kubernetes.io/projected/f33dc056-b642-4cf9-95d4-9c40ac0468ca-kube-api-access-csrhw\") pod \"horizon-operator-controller-manager-5d494799bf-lnvbp\" (UID: \"f33dc056-b642-4cf9-95d4-9c40ac0468ca\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.950021 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpcwp\" (UniqueName: \"kubernetes.io/projected/a836e80a-5070-4ac6-9d06-13f7dcb8f5a7-kube-api-access-qpcwp\") pod \"keystone-operator-controller-manager-7b4567c7cf-7tqrc\" (UID: \"a836e80a-5070-4ac6-9d06-13f7dcb8f5a7\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.950056 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xlxd\" (UniqueName: \"kubernetes.io/projected/04fbac3c-ab55-4790-85b2-3269762d6a55-kube-api-access-6xlxd\") pod \"glance-operator-controller-manager-589cbd6b5b-6h5f6\" (UID: \"04fbac3c-ab55-4790-85b2-3269762d6a55\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.950077 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:08 crc kubenswrapper[4862]: I1128 11:40:08.958508 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-k9kpx" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:08.985327 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.013518 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdtvr\" (UniqueName: \"kubernetes.io/projected/50269334-ed51-4bfc-8b87-1243d75774a3-kube-api-access-fdtvr\") pod \"heat-operator-controller-manager-5b77f656f-v5t45\" (UID: \"50269334-ed51-4bfc-8b87-1243d75774a3\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.026183 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.027799 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xlxd\" (UniqueName: \"kubernetes.io/projected/04fbac3c-ab55-4790-85b2-3269762d6a55-kube-api-access-6xlxd\") pod \"glance-operator-controller-manager-589cbd6b5b-6h5f6\" (UID: \"04fbac3c-ab55-4790-85b2-3269762d6a55\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.033021 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.034135 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.038747 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-9mpn5" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.042055 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.050650 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csrhw\" (UniqueName: \"kubernetes.io/projected/f33dc056-b642-4cf9-95d4-9c40ac0468ca-kube-api-access-csrhw\") pod \"horizon-operator-controller-manager-5d494799bf-lnvbp\" (UID: \"f33dc056-b642-4cf9-95d4-9c40ac0468ca\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.052058 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.056980 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.070303 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.071392 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.074008 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hhb9\" (UniqueName: \"kubernetes.io/projected/4ae22c3f-2ad0-4bde-8045-8e30d5d65d64-kube-api-access-9hhb9\") pod \"neutron-operator-controller-manager-6fdcddb789-776s4\" (UID: \"4ae22c3f-2ad0-4bde-8045-8e30d5d65d64\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.074053 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2lw2\" (UniqueName: \"kubernetes.io/projected/f4857ca4-a829-4205-b776-fa73faf12eac-kube-api-access-c2lw2\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.074144 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mwcs\" (UniqueName: \"kubernetes.io/projected/4c8abe61-f6dd-4bf2-80e6-c13db3db907f-kube-api-access-9mwcs\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-vhg8w\" (UID: \"4c8abe61-f6dd-4bf2-80e6-c13db3db907f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.074172 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpcwp\" (UniqueName: \"kubernetes.io/projected/a836e80a-5070-4ac6-9d06-13f7dcb8f5a7-kube-api-access-qpcwp\") pod \"keystone-operator-controller-manager-7b4567c7cf-7tqrc\" (UID: \"a836e80a-5070-4ac6-9d06-13f7dcb8f5a7\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.074199 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz6m7\" (UniqueName: \"kubernetes.io/projected/9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4-kube-api-access-fz6m7\") pod \"manila-operator-controller-manager-5d499bf58b-tmq8w\" (UID: \"9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.074224 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.074257 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhfwj\" (UniqueName: \"kubernetes.io/projected/c50b92c1-e947-4f41-abfe-278e1826a0c9-kube-api-access-jhfwj\") pod \"ironic-operator-controller-manager-67cb4dc6d4-dmqbn\" (UID: \"c50b92c1-e947-4f41-abfe-278e1826a0c9\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.074797 4862 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.074842 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert podName:f4857ca4-a829-4205-b776-fa73faf12eac nodeName:}" failed. No retries permitted until 2025-11-28 11:40:09.574825681 +0000 UTC m=+1087.107339602 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert") pod "infra-operator-controller-manager-57548d458d-nxpb6" (UID: "f4857ca4-a829-4205-b776-fa73faf12eac") : secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.075182 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rlhpv" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.076374 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.096586 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhfwj\" (UniqueName: \"kubernetes.io/projected/c50b92c1-e947-4f41-abfe-278e1826a0c9-kube-api-access-jhfwj\") pod \"ironic-operator-controller-manager-67cb4dc6d4-dmqbn\" (UID: \"c50b92c1-e947-4f41-abfe-278e1826a0c9\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.099290 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpcwp\" (UniqueName: \"kubernetes.io/projected/a836e80a-5070-4ac6-9d06-13f7dcb8f5a7-kube-api-access-qpcwp\") pod \"keystone-operator-controller-manager-7b4567c7cf-7tqrc\" (UID: \"a836e80a-5070-4ac6-9d06-13f7dcb8f5a7\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.102673 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2lw2\" (UniqueName: \"kubernetes.io/projected/f4857ca4-a829-4205-b776-fa73faf12eac-kube-api-access-c2lw2\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.112451 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.142183 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-54k27"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.143340 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.148006 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-54k27"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.156648 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-2jrnb" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.157929 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.159276 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.167026 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.167967 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.168003 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.169757 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-lzrkh" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.171610 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.171810 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-n8hbb" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.176872 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtkpv\" (UniqueName: \"kubernetes.io/projected/5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56-kube-api-access-xtkpv\") pod \"octavia-operator-controller-manager-64cdc6ff96-xsqj2\" (UID: \"5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.176929 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l4zs\" (UniqueName: \"kubernetes.io/projected/a0567ace-33d0-4e27-a4c6-739b24a847c8-kube-api-access-7l4zs\") pod \"nova-operator-controller-manager-79556f57fc-r6stc\" (UID: \"a0567ace-33d0-4e27-a4c6-739b24a847c8\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.176962 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hhb9\" (UniqueName: \"kubernetes.io/projected/4ae22c3f-2ad0-4bde-8045-8e30d5d65d64-kube-api-access-9hhb9\") pod \"neutron-operator-controller-manager-6fdcddb789-776s4\" (UID: \"4ae22c3f-2ad0-4bde-8045-8e30d5d65d64\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.177020 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mwcs\" (UniqueName: \"kubernetes.io/projected/4c8abe61-f6dd-4bf2-80e6-c13db3db907f-kube-api-access-9mwcs\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-vhg8w\" (UID: \"4c8abe61-f6dd-4bf2-80e6-c13db3db907f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.177058 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz6m7\" (UniqueName: \"kubernetes.io/projected/9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4-kube-api-access-fz6m7\") pod \"manila-operator-controller-manager-5d499bf58b-tmq8w\" (UID: \"9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.182158 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.195385 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.200844 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hhb9\" (UniqueName: \"kubernetes.io/projected/4ae22c3f-2ad0-4bde-8045-8e30d5d65d64-kube-api-access-9hhb9\") pod \"neutron-operator-controller-manager-6fdcddb789-776s4\" (UID: \"4ae22c3f-2ad0-4bde-8045-8e30d5d65d64\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.202327 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz6m7\" (UniqueName: \"kubernetes.io/projected/9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4-kube-api-access-fz6m7\") pod \"manila-operator-controller-manager-5d499bf58b-tmq8w\" (UID: \"9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.212714 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mwcs\" (UniqueName: \"kubernetes.io/projected/4c8abe61-f6dd-4bf2-80e6-c13db3db907f-kube-api-access-9mwcs\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-vhg8w\" (UID: \"4c8abe61-f6dd-4bf2-80e6-c13db3db907f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.236689 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.249409 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.255009 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.259357 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-zndns" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.276049 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.281950 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-729tk\" (UniqueName: \"kubernetes.io/projected/be84305c-9d9e-425d-a8d3-f1ce56254552-kube-api-access-729tk\") pod \"ovn-operator-controller-manager-56897c768d-54k27\" (UID: \"be84305c-9d9e-425d-a8d3-f1ce56254552\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.282044 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtkpv\" (UniqueName: \"kubernetes.io/projected/5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56-kube-api-access-xtkpv\") pod \"octavia-operator-controller-manager-64cdc6ff96-xsqj2\" (UID: \"5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.282084 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l4zs\" (UniqueName: \"kubernetes.io/projected/a0567ace-33d0-4e27-a4c6-739b24a847c8-kube-api-access-7l4zs\") pod \"nova-operator-controller-manager-79556f57fc-r6stc\" (UID: \"a0567ace-33d0-4e27-a4c6-739b24a847c8\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.282124 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.282178 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxfwb\" (UniqueName: \"kubernetes.io/projected/84002eaf-76bf-4a7e-a7ef-81c87d450270-kube-api-access-qxfwb\") pod \"placement-operator-controller-manager-57988cc5b5-n2hrf\" (UID: \"84002eaf-76bf-4a7e-a7ef-81c87d450270\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.282203 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nb4n\" (UniqueName: \"kubernetes.io/projected/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-kube-api-access-7nb4n\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.299662 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.302210 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtkpv\" (UniqueName: \"kubernetes.io/projected/5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56-kube-api-access-xtkpv\") pod \"octavia-operator-controller-manager-64cdc6ff96-xsqj2\" (UID: \"5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.303898 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l4zs\" (UniqueName: \"kubernetes.io/projected/a0567ace-33d0-4e27-a4c6-739b24a847c8-kube-api-access-7l4zs\") pod \"nova-operator-controller-manager-79556f57fc-r6stc\" (UID: \"a0567ace-33d0-4e27-a4c6-739b24a847c8\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.306538 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.327296 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.337919 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.339331 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.352504 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-9t8lb" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.375215 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.376284 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.399201 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-gshhw" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.401441 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.404427 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.404500 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxfwb\" (UniqueName: \"kubernetes.io/projected/84002eaf-76bf-4a7e-a7ef-81c87d450270-kube-api-access-qxfwb\") pod \"placement-operator-controller-manager-57988cc5b5-n2hrf\" (UID: \"84002eaf-76bf-4a7e-a7ef-81c87d450270\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.404536 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nb4n\" (UniqueName: \"kubernetes.io/projected/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-kube-api-access-7nb4n\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.404581 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-729tk\" (UniqueName: \"kubernetes.io/projected/be84305c-9d9e-425d-a8d3-f1ce56254552-kube-api-access-729tk\") pod \"ovn-operator-controller-manager-56897c768d-54k27\" (UID: \"be84305c-9d9e-425d-a8d3-f1ce56254552\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.404634 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zctk\" (UniqueName: \"kubernetes.io/projected/9e68a5c6-829c-4897-b251-c181097105fc-kube-api-access-8zctk\") pod \"swift-operator-controller-manager-d77b94747-jz7cq\" (UID: \"9e68a5c6-829c-4897-b251-c181097105fc\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.406172 4862 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.406253 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert podName:36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e nodeName:}" failed. No retries permitted until 2025-11-28 11:40:09.9062299 +0000 UTC m=+1087.438743821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" (UID: "36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.407307 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.438642 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.477276 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.489467 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.498007 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.512646 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.512646 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7btj\" (UniqueName: \"kubernetes.io/projected/176a3ccc-ec57-4af5-9e09-cb958ed71a2b-kube-api-access-l7btj\") pod \"test-operator-controller-manager-5cd6c7f4c8-d4bgt\" (UID: \"176a3ccc-ec57-4af5-9e09-cb958ed71a2b\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.512743 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msh5k\" (UniqueName: \"kubernetes.io/projected/70903410-bdee-42fc-a1bf-2213f259dd6f-kube-api-access-msh5k\") pod \"watcher-operator-controller-manager-656dcb59d4-p5n9j\" (UID: \"70903410-bdee-42fc-a1bf-2213f259dd6f\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.512764 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkddq\" (UniqueName: \"kubernetes.io/projected/d8d42319-792f-4aec-abcc-7158a1bf9f9c-kube-api-access-gkddq\") pod \"telemetry-operator-controller-manager-76cc84c6bb-trj7j\" (UID: \"d8d42319-792f-4aec-abcc-7158a1bf9f9c\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.512809 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zctk\" (UniqueName: \"kubernetes.io/projected/9e68a5c6-829c-4897-b251-c181097105fc-kube-api-access-8zctk\") pod \"swift-operator-controller-manager-d77b94747-jz7cq\" (UID: \"9e68a5c6-829c-4897-b251-c181097105fc\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.513462 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-729tk\" (UniqueName: \"kubernetes.io/projected/be84305c-9d9e-425d-a8d3-f1ce56254552-kube-api-access-729tk\") pod \"ovn-operator-controller-manager-56897c768d-54k27\" (UID: \"be84305c-9d9e-425d-a8d3-f1ce56254552\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.525434 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nb4n\" (UniqueName: \"kubernetes.io/projected/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-kube-api-access-7nb4n\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.525810 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-cpt8q" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.527034 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.535040 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.535972 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.545010 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxfwb\" (UniqueName: \"kubernetes.io/projected/84002eaf-76bf-4a7e-a7ef-81c87d450270-kube-api-access-qxfwb\") pod \"placement-operator-controller-manager-57988cc5b5-n2hrf\" (UID: \"84002eaf-76bf-4a7e-a7ef-81c87d450270\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.545226 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.545457 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.545710 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-8mqtp" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.556452 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.591943 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.593027 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.596139 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-t9cj7" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.602488 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.603928 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zctk\" (UniqueName: \"kubernetes.io/projected/9e68a5c6-829c-4897-b251-c181097105fc-kube-api-access-8zctk\") pod \"swift-operator-controller-manager-d77b94747-jz7cq\" (UID: \"9e68a5c6-829c-4897-b251-c181097105fc\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.613711 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.613744 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msh5k\" (UniqueName: \"kubernetes.io/projected/70903410-bdee-42fc-a1bf-2213f259dd6f-kube-api-access-msh5k\") pod \"watcher-operator-controller-manager-656dcb59d4-p5n9j\" (UID: \"70903410-bdee-42fc-a1bf-2213f259dd6f\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.613764 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkddq\" (UniqueName: \"kubernetes.io/projected/d8d42319-792f-4aec-abcc-7158a1bf9f9c-kube-api-access-gkddq\") pod \"telemetry-operator-controller-manager-76cc84c6bb-trj7j\" (UID: \"d8d42319-792f-4aec-abcc-7158a1bf9f9c\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.613853 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7btj\" (UniqueName: \"kubernetes.io/projected/176a3ccc-ec57-4af5-9e09-cb958ed71a2b-kube-api-access-l7btj\") pod \"test-operator-controller-manager-5cd6c7f4c8-d4bgt\" (UID: \"176a3ccc-ec57-4af5-9e09-cb958ed71a2b\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.614152 4862 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.614193 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert podName:f4857ca4-a829-4205-b776-fa73faf12eac nodeName:}" failed. No retries permitted until 2025-11-28 11:40:10.614179831 +0000 UTC m=+1088.146693752 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert") pod "infra-operator-controller-manager-57548d458d-nxpb6" (UID: "f4857ca4-a829-4205-b776-fa73faf12eac") : secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.618064 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.649826 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.661246 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7btj\" (UniqueName: \"kubernetes.io/projected/176a3ccc-ec57-4af5-9e09-cb958ed71a2b-kube-api-access-l7btj\") pod \"test-operator-controller-manager-5cd6c7f4c8-d4bgt\" (UID: \"176a3ccc-ec57-4af5-9e09-cb958ed71a2b\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.664009 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msh5k\" (UniqueName: \"kubernetes.io/projected/70903410-bdee-42fc-a1bf-2213f259dd6f-kube-api-access-msh5k\") pod \"watcher-operator-controller-manager-656dcb59d4-p5n9j\" (UID: \"70903410-bdee-42fc-a1bf-2213f259dd6f\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.706167 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.714041 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkddq\" (UniqueName: \"kubernetes.io/projected/d8d42319-792f-4aec-abcc-7158a1bf9f9c-kube-api-access-gkddq\") pod \"telemetry-operator-controller-manager-76cc84c6bb-trj7j\" (UID: \"d8d42319-792f-4aec-abcc-7158a1bf9f9c\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.715380 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.716511 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wfxg\" (UniqueName: \"kubernetes.io/projected/50d316aa-a1e4-4ca0-907a-52537f1b1218-kube-api-access-5wfxg\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xtc9t\" (UID: \"50d316aa-a1e4-4ca0-907a-52537f1b1218\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.716620 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g299\" (UniqueName: \"kubernetes.io/projected/b19e018e-f54a-4f60-bef9-345839ae970c-kube-api-access-7g299\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.716656 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.818270 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.818545 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wfxg\" (UniqueName: \"kubernetes.io/projected/50d316aa-a1e4-4ca0-907a-52537f1b1218-kube-api-access-5wfxg\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xtc9t\" (UID: \"50d316aa-a1e4-4ca0-907a-52537f1b1218\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.818575 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g299\" (UniqueName: \"kubernetes.io/projected/b19e018e-f54a-4f60-bef9-345839ae970c-kube-api-access-7g299\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.818639 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.818874 4862 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.818944 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:10.31892792 +0000 UTC m=+1087.851441841 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "metrics-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.819053 4862 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.819117 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:10.319101105 +0000 UTC m=+1087.851615026 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.827258 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.849950 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g299\" (UniqueName: \"kubernetes.io/projected/b19e018e-f54a-4f60-bef9-345839ae970c-kube-api-access-7g299\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.850784 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wfxg\" (UniqueName: \"kubernetes.io/projected/50d316aa-a1e4-4ca0-907a-52537f1b1218-kube-api-access-5wfxg\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xtc9t\" (UID: \"50d316aa-a1e4-4ca0-907a-52537f1b1218\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.855974 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.892751 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.930243 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.930441 4862 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: E1128 11:40:09.930494 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert podName:36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e nodeName:}" failed. No retries permitted until 2025-11-28 11:40:10.930477685 +0000 UTC m=+1088.462991606 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" (UID: "36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.933648 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh"] Nov 28 11:40:09 crc kubenswrapper[4862]: I1128 11:40:09.970071 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.179786 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.180702 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" event={"ID":"f8ce3d3e-999e-48bf-887a-add2b478966d","Type":"ContainerStarted","Data":"f1e8833a3b6a70a02e8b368bc149ffb0a3f578b2cbc35d8e5814bca85b8ba87f"} Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.200773 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-5qbwv"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.341110 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.341174 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.341285 4862 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.341331 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:11.341316092 +0000 UTC m=+1088.873830013 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "metrics-server-cert" not found Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.342416 4862 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.342448 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:11.342439004 +0000 UTC m=+1088.874952915 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "webhook-server-cert" not found Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.450498 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.465242 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.471028 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc"] Nov 28 11:40:10 crc kubenswrapper[4862]: W1128 11:40:10.483369 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04fbac3c_ab55_4790_85b2_3269762d6a55.slice/crio-db9ec96023c5db24de2a521d81636b16b1784f807630adf6c3f9a2a3a52cda2a WatchSource:0}: Error finding container db9ec96023c5db24de2a521d81636b16b1784f807630adf6c3f9a2a3a52cda2a: Status 404 returned error can't find the container with id db9ec96023c5db24de2a521d81636b16b1784f807630adf6c3f9a2a3a52cda2a Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.644498 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.646847 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.647136 4862 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.647198 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert podName:f4857ca4-a829-4205-b776-fa73faf12eac nodeName:}" failed. No retries permitted until 2025-11-28 11:40:12.647167604 +0000 UTC m=+1090.179681525 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert") pod "infra-operator-controller-manager-57548d458d-nxpb6" (UID: "f4857ca4-a829-4205-b776-fa73faf12eac") : secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.663157 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp"] Nov 28 11:40:10 crc kubenswrapper[4862]: W1128 11:40:10.669262 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf33dc056_b642_4cf9_95d4_9c40ac0468ca.slice/crio-3ca5e1ee194051cedde05791f4e0894ead433d92a22601596d72cb310387abcb WatchSource:0}: Error finding container 3ca5e1ee194051cedde05791f4e0894ead433d92a22601596d72cb310387abcb: Status 404 returned error can't find the container with id 3ca5e1ee194051cedde05791f4e0894ead433d92a22601596d72cb310387abcb Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.669991 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.777853 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.785628 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.792681 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.912338 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.932999 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.946121 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-54k27"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.951052 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.952166 4862 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.952213 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert podName:36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e nodeName:}" failed. No retries permitted until 2025-11-28 11:40:12.952197402 +0000 UTC m=+1090.484711413 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" (UID: "36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.955412 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.960591 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j"] Nov 28 11:40:10 crc kubenswrapper[4862]: W1128 11:40:10.971552 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8d42319_792f_4aec_abcc_7158a1bf9f9c.slice/crio-b4d362f86e4f415bb46d39760a5b76a6f55932d0c5b879d493d19a89732ebc28 WatchSource:0}: Error finding container b4d362f86e4f415bb46d39760a5b76a6f55932d0c5b879d493d19a89732ebc28: Status 404 returned error can't find the container with id b4d362f86e4f415bb46d39760a5b76a6f55932d0c5b879d493d19a89732ebc28 Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.971591 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt"] Nov 28 11:40:10 crc kubenswrapper[4862]: I1128 11:40:10.979339 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf"] Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.980567 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gkddq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-trj7j_openstack-operators(d8d42319-792f-4aec-abcc-7158a1bf9f9c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.980675 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7l4zs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-r6stc_openstack-operators(a0567ace-33d0-4e27-a4c6-739b24a847c8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.980800 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9hhb9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6fdcddb789-776s4_openstack-operators(4ae22c3f-2ad0-4bde-8045-8e30d5d65d64): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.981311 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l7btj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-d4bgt_openstack-operators(176a3ccc-ec57-4af5-9e09-cb958ed71a2b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.983304 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7l4zs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-r6stc_openstack-operators(a0567ace-33d0-4e27-a4c6-739b24a847c8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.983624 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9hhb9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6fdcddb789-776s4_openstack-operators(4ae22c3f-2ad0-4bde-8045-8e30d5d65d64): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.983714 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gkddq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-trj7j_openstack-operators(d8d42319-792f-4aec-abcc-7158a1bf9f9c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.983401 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l7btj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-d4bgt_openstack-operators(176a3ccc-ec57-4af5-9e09-cb958ed71a2b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.984681 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" podUID="4ae22c3f-2ad0-4bde-8045-8e30d5d65d64" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.984683 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" podUID="a0567ace-33d0-4e27-a4c6-739b24a847c8" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.985117 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" podUID="176a3ccc-ec57-4af5-9e09-cb958ed71a2b" Nov 28 11:40:10 crc kubenswrapper[4862]: E1128 11:40:10.985146 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" podUID="d8d42319-792f-4aec-abcc-7158a1bf9f9c" Nov 28 11:40:10 crc kubenswrapper[4862]: W1128 11:40:10.998837 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84002eaf_76bf_4a7e_a7ef_81c87d450270.slice/crio-cb651c005343dbbef184d400b5b66766c3b1e931a4674ecb6054702e7da7722e WatchSource:0}: Error finding container cb651c005343dbbef184d400b5b66766c3b1e931a4674ecb6054702e7da7722e: Status 404 returned error can't find the container with id cb651c005343dbbef184d400b5b66766c3b1e931a4674ecb6054702e7da7722e Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.004632 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qxfwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-n2hrf_openstack-operators(84002eaf-76bf-4a7e-a7ef-81c87d450270): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.008762 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qxfwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-n2hrf_openstack-operators(84002eaf-76bf-4a7e-a7ef-81c87d450270): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.009896 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" podUID="84002eaf-76bf-4a7e-a7ef-81c87d450270" Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.049074 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t"] Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.059937 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5wfxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-xtc9t_openstack-operators(50d316aa-a1e4-4ca0-907a-52537f1b1218): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.061135 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" podUID="50d316aa-a1e4-4ca0-907a-52537f1b1218" Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.210129 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" event={"ID":"a0567ace-33d0-4e27-a4c6-739b24a847c8","Type":"ContainerStarted","Data":"5f4415c87b3edaa3bd4570b37464f4ed9f3aef99111ff5645426354f667cf26f"} Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.211842 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" podUID="a0567ace-33d0-4e27-a4c6-739b24a847c8" Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.212395 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" event={"ID":"176a3ccc-ec57-4af5-9e09-cb958ed71a2b","Type":"ContainerStarted","Data":"efe9ec1e3e15a38f986244fce1c4a76910acb5060bf9b2f7480bd9f43c2e4c5f"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.213412 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" event={"ID":"b1165cf7-ed1f-48ba-a6dc-1cab53edf166","Type":"ContainerStarted","Data":"f6fb2f4c205881d68ede241d60bc7ff24926357b27e8205c26c89c5e8af0e8f8"} Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.214550 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" podUID="176a3ccc-ec57-4af5-9e09-cb958ed71a2b" Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.215184 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" event={"ID":"be84305c-9d9e-425d-a8d3-f1ce56254552","Type":"ContainerStarted","Data":"15946f74a210ef6374e50f62d33663abe9c1362827d6b37f9aec6254cda6a035"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.217180 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" event={"ID":"04fbac3c-ab55-4790-85b2-3269762d6a55","Type":"ContainerStarted","Data":"db9ec96023c5db24de2a521d81636b16b1784f807630adf6c3f9a2a3a52cda2a"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.218234 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" event={"ID":"4ae22c3f-2ad0-4bde-8045-8e30d5d65d64","Type":"ContainerStarted","Data":"bd7a1ef2ca8ed6e25310cecfe7019f2f9f559c0c58902e7e0a9ee19572fedcde"} Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.221207 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" podUID="4ae22c3f-2ad0-4bde-8045-8e30d5d65d64" Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.221766 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" event={"ID":"50d316aa-a1e4-4ca0-907a-52537f1b1218","Type":"ContainerStarted","Data":"6cb73378b868754d6b922547e267cfeef235c8944eb7ce1c9b4fa9b7e5fead56"} Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.223618 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" podUID="50d316aa-a1e4-4ca0-907a-52537f1b1218" Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.229614 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" event={"ID":"4c8abe61-f6dd-4bf2-80e6-c13db3db907f","Type":"ContainerStarted","Data":"8248464e2b9b53a5c42af7e5a4864a0ac0461b9c562bbd2c0220843c0d8e3a98"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.233667 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" event={"ID":"9e68a5c6-829c-4897-b251-c181097105fc","Type":"ContainerStarted","Data":"8d4d4dcfdd8cf7f00ea41e4451b9d505f7184915e02d70d9235b282d21ab0706"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.254746 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" event={"ID":"5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56","Type":"ContainerStarted","Data":"69db421a84a666440f6a68dc2b91d2e4b93aea44af7fb0827d48f2e91ae9d5f9"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.267321 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" event={"ID":"84002eaf-76bf-4a7e-a7ef-81c87d450270","Type":"ContainerStarted","Data":"cb651c005343dbbef184d400b5b66766c3b1e931a4674ecb6054702e7da7722e"} Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.272451 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" podUID="84002eaf-76bf-4a7e-a7ef-81c87d450270" Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.272941 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" event={"ID":"70903410-bdee-42fc-a1bf-2213f259dd6f","Type":"ContainerStarted","Data":"e3db981afe30c86ba0c045b69efc58a56a7d8e58568d016395c4b2cd5a2187b2"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.276019 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" event={"ID":"8bba217a-0d90-4544-b772-62e6e5e58d74","Type":"ContainerStarted","Data":"166e2cb8e7a405c908eefd2834f73336180b35783606005a5b4adfe3c015e0aa"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.281905 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" event={"ID":"a836e80a-5070-4ac6-9d06-13f7dcb8f5a7","Type":"ContainerStarted","Data":"da2a1a51ebf2f0b902ce686eeeebafd432300a7712936b51937c20a672038de8"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.285979 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" event={"ID":"c50b92c1-e947-4f41-abfe-278e1826a0c9","Type":"ContainerStarted","Data":"d5a531e5d79a78578f113c75a705f2631d8cba46a6c142fbc9783212f23f1aad"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.296293 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" event={"ID":"f33dc056-b642-4cf9-95d4-9c40ac0468ca","Type":"ContainerStarted","Data":"3ca5e1ee194051cedde05791f4e0894ead433d92a22601596d72cb310387abcb"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.301166 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" event={"ID":"d8d42319-792f-4aec-abcc-7158a1bf9f9c","Type":"ContainerStarted","Data":"b4d362f86e4f415bb46d39760a5b76a6f55932d0c5b879d493d19a89732ebc28"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.304394 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" event={"ID":"9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4","Type":"ContainerStarted","Data":"daaa2377e4aa8f786bdcf28322b8db4af7d800cb63ea7017fbd511e5393602a4"} Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.304745 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" podUID="d8d42319-792f-4aec-abcc-7158a1bf9f9c" Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.333473 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" event={"ID":"50269334-ed51-4bfc-8b87-1243d75774a3","Type":"ContainerStarted","Data":"c5a516efefac8c461d73e59f30040a8b8b9aa75a0f844eba473aa630acc5da49"} Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.356869 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:11 crc kubenswrapper[4862]: I1128 11:40:11.357018 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.357174 4862 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.357228 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:13.357210471 +0000 UTC m=+1090.889724392 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "metrics-server-cert" not found Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.358282 4862 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 11:40:11 crc kubenswrapper[4862]: E1128 11:40:11.358374 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:13.358346904 +0000 UTC m=+1090.890860875 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "webhook-server-cert" not found Nov 28 11:40:12 crc kubenswrapper[4862]: E1128 11:40:12.351699 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" podUID="50d316aa-a1e4-4ca0-907a-52537f1b1218" Nov 28 11:40:12 crc kubenswrapper[4862]: E1128 11:40:12.352625 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" podUID="d8d42319-792f-4aec-abcc-7158a1bf9f9c" Nov 28 11:40:12 crc kubenswrapper[4862]: E1128 11:40:12.357428 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" podUID="84002eaf-76bf-4a7e-a7ef-81c87d450270" Nov 28 11:40:12 crc kubenswrapper[4862]: E1128 11:40:12.357511 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" podUID="176a3ccc-ec57-4af5-9e09-cb958ed71a2b" Nov 28 11:40:12 crc kubenswrapper[4862]: E1128 11:40:12.357556 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" podUID="4ae22c3f-2ad0-4bde-8045-8e30d5d65d64" Nov 28 11:40:12 crc kubenswrapper[4862]: E1128 11:40:12.363404 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" podUID="a0567ace-33d0-4e27-a4c6-739b24a847c8" Nov 28 11:40:12 crc kubenswrapper[4862]: I1128 11:40:12.703859 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:12 crc kubenswrapper[4862]: E1128 11:40:12.704119 4862 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:12 crc kubenswrapper[4862]: E1128 11:40:12.704199 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert podName:f4857ca4-a829-4205-b776-fa73faf12eac nodeName:}" failed. No retries permitted until 2025-11-28 11:40:16.70418045 +0000 UTC m=+1094.236694371 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert") pod "infra-operator-controller-manager-57548d458d-nxpb6" (UID: "f4857ca4-a829-4205-b776-fa73faf12eac") : secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:13 crc kubenswrapper[4862]: I1128 11:40:13.008330 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:13 crc kubenswrapper[4862]: E1128 11:40:13.008543 4862 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:13 crc kubenswrapper[4862]: E1128 11:40:13.008613 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert podName:36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e nodeName:}" failed. No retries permitted until 2025-11-28 11:40:17.008595741 +0000 UTC m=+1094.541109662 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" (UID: "36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:13 crc kubenswrapper[4862]: I1128 11:40:13.414038 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:13 crc kubenswrapper[4862]: I1128 11:40:13.414144 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:13 crc kubenswrapper[4862]: E1128 11:40:13.414261 4862 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 11:40:13 crc kubenswrapper[4862]: E1128 11:40:13.414301 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:17.41428928 +0000 UTC m=+1094.946803201 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "metrics-server-cert" not found Nov 28 11:40:13 crc kubenswrapper[4862]: E1128 11:40:13.414627 4862 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 11:40:13 crc kubenswrapper[4862]: E1128 11:40:13.414652 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:17.41464367 +0000 UTC m=+1094.947157591 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "webhook-server-cert" not found Nov 28 11:40:16 crc kubenswrapper[4862]: I1128 11:40:16.796853 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:16 crc kubenswrapper[4862]: E1128 11:40:16.797081 4862 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:16 crc kubenswrapper[4862]: E1128 11:40:16.797353 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert podName:f4857ca4-a829-4205-b776-fa73faf12eac nodeName:}" failed. No retries permitted until 2025-11-28 11:40:24.797319852 +0000 UTC m=+1102.329833773 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert") pod "infra-operator-controller-manager-57548d458d-nxpb6" (UID: "f4857ca4-a829-4205-b776-fa73faf12eac") : secret "infra-operator-webhook-server-cert" not found Nov 28 11:40:17 crc kubenswrapper[4862]: I1128 11:40:17.101553 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:17 crc kubenswrapper[4862]: E1128 11:40:17.101753 4862 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:17 crc kubenswrapper[4862]: E1128 11:40:17.102162 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert podName:36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e nodeName:}" failed. No retries permitted until 2025-11-28 11:40:25.102143205 +0000 UTC m=+1102.634657126 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" (UID: "36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 11:40:17 crc kubenswrapper[4862]: I1128 11:40:17.506524 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:17 crc kubenswrapper[4862]: I1128 11:40:17.506693 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:17 crc kubenswrapper[4862]: E1128 11:40:17.506734 4862 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 11:40:17 crc kubenswrapper[4862]: E1128 11:40:17.506819 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:25.506794944 +0000 UTC m=+1103.039308945 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "metrics-server-cert" not found Nov 28 11:40:17 crc kubenswrapper[4862]: E1128 11:40:17.506860 4862 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 11:40:17 crc kubenswrapper[4862]: E1128 11:40:17.506943 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs podName:b19e018e-f54a-4f60-bef9-345839ae970c nodeName:}" failed. No retries permitted until 2025-11-28 11:40:25.506928528 +0000 UTC m=+1103.039442519 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-zf5gf" (UID: "b19e018e-f54a-4f60-bef9-345839ae970c") : secret "webhook-server-cert" not found Nov 28 11:40:24 crc kubenswrapper[4862]: I1128 11:40:24.835135 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:24 crc kubenswrapper[4862]: I1128 11:40:24.842813 4862 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 11:40:24 crc kubenswrapper[4862]: I1128 11:40:24.859496 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4857ca4-a829-4205-b776-fa73faf12eac-cert\") pod \"infra-operator-controller-manager-57548d458d-nxpb6\" (UID: \"f4857ca4-a829-4205-b776-fa73faf12eac\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:25 crc kubenswrapper[4862]: I1128 11:40:25.040446 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:25 crc kubenswrapper[4862]: I1128 11:40:25.140704 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:25 crc kubenswrapper[4862]: I1128 11:40:25.151693 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67glltc\" (UID: \"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:25 crc kubenswrapper[4862]: I1128 11:40:25.378736 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:25 crc kubenswrapper[4862]: I1128 11:40:25.546986 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:25 crc kubenswrapper[4862]: I1128 11:40:25.547268 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:25 crc kubenswrapper[4862]: I1128 11:40:25.551606 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:25 crc kubenswrapper[4862]: I1128 11:40:25.554864 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b19e018e-f54a-4f60-bef9-345839ae970c-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-zf5gf\" (UID: \"b19e018e-f54a-4f60-bef9-345839ae970c\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:25 crc kubenswrapper[4862]: I1128 11:40:25.799306 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:26 crc kubenswrapper[4862]: E1128 11:40:26.088284 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:2ee37ff474bee3203447df4f326a9279a515e770573153338296dd074722c677" Nov 28 11:40:26 crc kubenswrapper[4862]: E1128 11:40:26.088467 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:2ee37ff474bee3203447df4f326a9279a515e770573153338296dd074722c677,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fdtvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5b77f656f-v5t45_openstack-operators(50269334-ed51-4bfc-8b87-1243d75774a3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:40:27 crc kubenswrapper[4862]: E1128 11:40:27.945260 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:888edf6f432e52eaa5fc3caeae616fe38a3302b006bbba0e38885b2beba9f0f2" Nov 28 11:40:27 crc kubenswrapper[4862]: E1128 11:40:27.945652 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:888edf6f432e52eaa5fc3caeae616fe38a3302b006bbba0e38885b2beba9f0f2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9mwcs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-66f4dd4bc7-vhg8w_openstack-operators(4c8abe61-f6dd-4bf2-80e6-c13db3db907f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:40:29 crc kubenswrapper[4862]: E1128 11:40:29.017368 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711" Nov 28 11:40:29 crc kubenswrapper[4862]: E1128 11:40:29.017755 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qpcwp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7b4567c7cf-7tqrc_openstack-operators(a836e80a-5070-4ac6-9d06-13f7dcb8f5a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:40:43 crc kubenswrapper[4862]: E1128 11:40:43.220830 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423" Nov 28 11:40:43 crc kubenswrapper[4862]: E1128 11:40:43.221693 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qxfwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-n2hrf_openstack-operators(84002eaf-76bf-4a7e-a7ef-81c87d450270): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:40:44 crc kubenswrapper[4862]: E1128 11:40:44.155714 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa" Nov 28 11:40:44 crc kubenswrapper[4862]: E1128 11:40:44.156196 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l7btj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-d4bgt_openstack-operators(176a3ccc-ec57-4af5-9e09-cb958ed71a2b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:40:44 crc kubenswrapper[4862]: I1128 11:40:44.228553 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6"] Nov 28 11:40:45 crc kubenswrapper[4862]: E1128 11:40:45.106579 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7" Nov 28 11:40:45 crc kubenswrapper[4862]: E1128 11:40:45.106827 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7l4zs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-r6stc_openstack-operators(a0567ace-33d0-4e27-a4c6-739b24a847c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:40:46 crc kubenswrapper[4862]: E1128 11:40:46.710218 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Nov 28 11:40:46 crc kubenswrapper[4862]: E1128 11:40:46.710655 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gkddq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-trj7j_openstack-operators(d8d42319-792f-4aec-abcc-7158a1bf9f9c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:40:47 crc kubenswrapper[4862]: I1128 11:40:47.667772 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc"] Nov 28 11:40:47 crc kubenswrapper[4862]: I1128 11:40:47.679603 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" event={"ID":"f4857ca4-a829-4205-b776-fa73faf12eac","Type":"ContainerStarted","Data":"8643a1d750743d3b1999e5806703f168e09f5f2bdb203f034d9dbed47f94ceff"} Nov 28 11:40:47 crc kubenswrapper[4862]: I1128 11:40:47.696583 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf"] Nov 28 11:40:47 crc kubenswrapper[4862]: E1128 11:40:47.873435 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 28 11:40:47 crc kubenswrapper[4862]: E1128 11:40:47.873633 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5wfxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-xtc9t_openstack-operators(50d316aa-a1e4-4ca0-907a-52537f1b1218): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:40:47 crc kubenswrapper[4862]: E1128 11:40:47.874971 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" podUID="50d316aa-a1e4-4ca0-907a-52537f1b1218" Nov 28 11:40:48 crc kubenswrapper[4862]: W1128 11:40:48.102457 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb19e018e_f54a_4f60_bef9_345839ae970c.slice/crio-dfd9258bac5c10dfe4789a63ca4df98c31f7ecca06f2e9ae12ffcf05ee98cb5e WatchSource:0}: Error finding container dfd9258bac5c10dfe4789a63ca4df98c31f7ecca06f2e9ae12ffcf05ee98cb5e: Status 404 returned error can't find the container with id dfd9258bac5c10dfe4789a63ca4df98c31f7ecca06f2e9ae12ffcf05ee98cb5e Nov 28 11:40:48 crc kubenswrapper[4862]: W1128 11:40:48.128579 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36af3eb7_8fd0_41b5_a7f7_8b42ecdecb8e.slice/crio-1e9e158e5f19fb910ddd2c46dcdffdb4dc571fe6dbbd1ca39dc45cf2a1fd7f22 WatchSource:0}: Error finding container 1e9e158e5f19fb910ddd2c46dcdffdb4dc571fe6dbbd1ca39dc45cf2a1fd7f22: Status 404 returned error can't find the container with id 1e9e158e5f19fb910ddd2c46dcdffdb4dc571fe6dbbd1ca39dc45cf2a1fd7f22 Nov 28 11:40:48 crc kubenswrapper[4862]: I1128 11:40:48.693865 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" event={"ID":"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e","Type":"ContainerStarted","Data":"1e9e158e5f19fb910ddd2c46dcdffdb4dc571fe6dbbd1ca39dc45cf2a1fd7f22"} Nov 28 11:40:48 crc kubenswrapper[4862]: I1128 11:40:48.709746 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" event={"ID":"b1165cf7-ed1f-48ba-a6dc-1cab53edf166","Type":"ContainerStarted","Data":"5e3ecfbbbd18685ce940a77b489cd4df4f909761c0dc25266a5a4d91532d0b0e"} Nov 28 11:40:48 crc kubenswrapper[4862]: I1128 11:40:48.713259 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" event={"ID":"be84305c-9d9e-425d-a8d3-f1ce56254552","Type":"ContainerStarted","Data":"060bebc27fae18bf72d61c35d0a3f4c6dcfc5d23fee8a955358beee4a3e6a8ed"} Nov 28 11:40:48 crc kubenswrapper[4862]: I1128 11:40:48.719372 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" event={"ID":"b19e018e-f54a-4f60-bef9-345839ae970c","Type":"ContainerStarted","Data":"dfd9258bac5c10dfe4789a63ca4df98c31f7ecca06f2e9ae12ffcf05ee98cb5e"} Nov 28 11:40:48 crc kubenswrapper[4862]: E1128 11:40:48.754030 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jhfwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-67cb4dc6d4-dmqbn_openstack-operators(c50b92c1-e947-4f41-abfe-278e1826a0c9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 11:40:48 crc kubenswrapper[4862]: E1128 11:40:48.755193 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" podUID="c50b92c1-e947-4f41-abfe-278e1826a0c9" Nov 28 11:40:49 crc kubenswrapper[4862]: E1128 11:40:49.449037 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 28 11:40:49 crc kubenswrapper[4862]: E1128 11:40:49.449210 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9mwcs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-66f4dd4bc7-vhg8w_openstack-operators(4c8abe61-f6dd-4bf2-80e6-c13db3db907f): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Nov 28 11:40:49 crc kubenswrapper[4862]: E1128 11:40:49.450570 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" podUID="4c8abe61-f6dd-4bf2-80e6-c13db3db907f" Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.735613 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" event={"ID":"b19e018e-f54a-4f60-bef9-345839ae970c","Type":"ContainerStarted","Data":"d3192e770f5112297c80c42a4e10c27452f4cb8ac7584c13cdd75f85ffea7cf4"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.735905 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.740357 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" event={"ID":"5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56","Type":"ContainerStarted","Data":"c686d4239ac5a1345868f4c224c0f84d870037a598835dd0dbdc152f8686f45e"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.741689 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" event={"ID":"9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4","Type":"ContainerStarted","Data":"c2d2e5905aab0694f2b5e7a53ba4d10c97d8cc62c471932e9e70baa4143074bd"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.742716 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" event={"ID":"8bba217a-0d90-4544-b772-62e6e5e58d74","Type":"ContainerStarted","Data":"1eec3c86d3916a594667df123a0f94afd7067c970d00ded9d6b5c8f15c1f3424"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.743692 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" event={"ID":"9e68a5c6-829c-4897-b251-c181097105fc","Type":"ContainerStarted","Data":"014390d1c08f0d8e593e62b014d410923b4d99fb8cd47482a95e2d0ff00183a4"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.744838 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" event={"ID":"c50b92c1-e947-4f41-abfe-278e1826a0c9","Type":"ContainerStarted","Data":"b5ed564eae349a35665334524dc7318ee11e2f10acf6f32281eecd63417a62f4"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.745462 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" Nov 28 11:40:49 crc kubenswrapper[4862]: E1128 11:40:49.751408 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" podUID="c50b92c1-e947-4f41-abfe-278e1826a0c9" Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.759008 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" event={"ID":"04fbac3c-ab55-4790-85b2-3269762d6a55","Type":"ContainerStarted","Data":"6eccde9d810dc0ccab5c6a7a5322b37d185d94b64648b29bd0f6e7d988803fc0"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.759986 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" event={"ID":"70903410-bdee-42fc-a1bf-2213f259dd6f","Type":"ContainerStarted","Data":"dedd8a47753f4ec0a7704eca662a252fde36e355fbd1374e814b2f00d81995cd"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.766417 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" event={"ID":"f33dc056-b642-4cf9-95d4-9c40ac0468ca","Type":"ContainerStarted","Data":"413ab228a9ea16677d8096d88cf4e00602af4887e37e132d018cc96508f7a76f"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.785358 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" event={"ID":"4ae22c3f-2ad0-4bde-8045-8e30d5d65d64","Type":"ContainerStarted","Data":"0be0f9df7f6ba9d00394052b072afdc6df65522523f4819270b4687c6972c591"} Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.790105 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" event={"ID":"f8ce3d3e-999e-48bf-887a-add2b478966d","Type":"ContainerStarted","Data":"bf4a93560e53b501766695fb54248c9705e11de4ad40193bdb55124deb9c8156"} Nov 28 11:40:49 crc kubenswrapper[4862]: E1128 11:40:49.824411 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 28 11:40:49 crc kubenswrapper[4862]: E1128 11:40:49.824552 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fdtvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5b77f656f-v5t45_openstack-operators(50269334-ed51-4bfc-8b87-1243d75774a3): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Nov 28 11:40:49 crc kubenswrapper[4862]: E1128 11:40:49.832336 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" podUID="50269334-ed51-4bfc-8b87-1243d75774a3" Nov 28 11:40:49 crc kubenswrapper[4862]: I1128 11:40:49.834592 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" podStartSLOduration=40.834580401 podStartE2EDuration="40.834580401s" podCreationTimestamp="2025-11-28 11:40:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:40:49.796565655 +0000 UTC m=+1127.329079576" watchObservedRunningTime="2025-11-28 11:40:49.834580401 +0000 UTC m=+1127.367094322" Nov 28 11:40:50 crc kubenswrapper[4862]: E1128 11:40:50.806888 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" podUID="c50b92c1-e947-4f41-abfe-278e1826a0c9" Nov 28 11:40:55 crc kubenswrapper[4862]: I1128 11:40:55.805569 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-zf5gf" Nov 28 11:40:55 crc kubenswrapper[4862]: I1128 11:40:55.860806 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" event={"ID":"50269334-ed51-4bfc-8b87-1243d75774a3","Type":"ContainerStarted","Data":"8335adfb49da3c30b24b55f680ac08501a063ad6b4e66f43d77435652c9feb9d"} Nov 28 11:40:57 crc kubenswrapper[4862]: E1128 11:40:57.366392 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" podUID="176a3ccc-ec57-4af5-9e09-cb958ed71a2b" Nov 28 11:40:57 crc kubenswrapper[4862]: E1128 11:40:57.660922 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" podUID="a0567ace-33d0-4e27-a4c6-739b24a847c8" Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.876716 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" event={"ID":"be84305c-9d9e-425d-a8d3-f1ce56254552","Type":"ContainerStarted","Data":"74bfd57c3be91478e8b925af11a7d032a6172daf9c87b8eb7378ba4169f0d2a1"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.878747 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" event={"ID":"f33dc056-b642-4cf9-95d4-9c40ac0468ca","Type":"ContainerStarted","Data":"57672acd93ca355c9f046ff067a3f88dc87e4529e556653279805a299b1182ba"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.880169 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" event={"ID":"d8d42319-792f-4aec-abcc-7158a1bf9f9c","Type":"ContainerStarted","Data":"838c3f60ab90d7dc822283958038f7b16ef6d821d2c9df4bc51d2ce6f0ca27d7"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.881526 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" event={"ID":"176a3ccc-ec57-4af5-9e09-cb958ed71a2b","Type":"ContainerStarted","Data":"11b350e101f97f9597123d0c856c484f3cd7b25835b54128eb06c7ebe5eb77b6"} Nov 28 11:40:57 crc kubenswrapper[4862]: E1128 11:40:57.882724 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" podUID="176a3ccc-ec57-4af5-9e09-cb958ed71a2b" Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.882975 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" event={"ID":"f8ce3d3e-999e-48bf-887a-add2b478966d","Type":"ContainerStarted","Data":"81b7bcfe1ced16c118dace014ef2fe89b66ddd84af7fcc095cc0aec4300b7d32"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.884109 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" event={"ID":"4c8abe61-f6dd-4bf2-80e6-c13db3db907f","Type":"ContainerStarted","Data":"3d11250a7c27b856a67fabcf13bed43c66e8073040095a9d9497c568cf334832"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.885671 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" event={"ID":"b1165cf7-ed1f-48ba-a6dc-1cab53edf166","Type":"ContainerStarted","Data":"4bd76c570fd7b85ef249c111204a4eecf2756274b3e54ac0f1c2f90aef6b0f6f"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.885901 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.888138 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" event={"ID":"70903410-bdee-42fc-a1bf-2213f259dd6f","Type":"ContainerStarted","Data":"17a311273da6e9c50074cfd3876dd2ebe4be73eed2445d0d6a1c77fa7e60e2b8"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.889406 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.889788 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" event={"ID":"5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56","Type":"ContainerStarted","Data":"4065d87c8211b463c4c6e008715a07a67605080af81a1b515eed295bb1220024"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.891069 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" event={"ID":"a0567ace-33d0-4e27-a4c6-739b24a847c8","Type":"ContainerStarted","Data":"d28145cf278f2a53fe3643e51631bb3e59e901d4bfcfce409aec75eb06ef10b6"} Nov 28 11:40:57 crc kubenswrapper[4862]: E1128 11:40:57.892279 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" podUID="a0567ace-33d0-4e27-a4c6-739b24a847c8" Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.892876 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" event={"ID":"84002eaf-76bf-4a7e-a7ef-81c87d450270","Type":"ContainerStarted","Data":"84bbe8750f0f17d5ed1fcd12f6c8b225b8204faa2c024e2791cd82a1a3cdf345"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.894198 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" event={"ID":"8bba217a-0d90-4544-b772-62e6e5e58d74","Type":"ContainerStarted","Data":"95ea60d0710fcd0f977c88dab2122380cd54576e36bb4d625dc77770e8faed78"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.895151 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" event={"ID":"f4857ca4-a829-4205-b776-fa73faf12eac","Type":"ContainerStarted","Data":"b2cc02cdbca4367fa47f4d80bdba5c321f45e11140dd4e00279da968ec743934"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.896182 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" event={"ID":"4ae22c3f-2ad0-4bde-8045-8e30d5d65d64","Type":"ContainerStarted","Data":"f749a983b6ee3beefa7d3bb12f8e7319a4ad751b67b6cb3e05e71ef1d788c078"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.897357 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" event={"ID":"9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4","Type":"ContainerStarted","Data":"a7041021cfd940c9a3fbb4b2f7a4c30341290b87419ba52518df4bc5bab7dbbe"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.898710 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" event={"ID":"50269334-ed51-4bfc-8b87-1243d75774a3","Type":"ContainerStarted","Data":"4633870e40d78dff3a82e87c8042a889e274b3f907aff2ffa2f028a174c9bd55"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.899589 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" event={"ID":"a836e80a-5070-4ac6-9d06-13f7dcb8f5a7","Type":"ContainerStarted","Data":"ae40b942a8717651ff52268d113945745deaf23153c59d810e11b6c365ff5076"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.922967 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" event={"ID":"04fbac3c-ab55-4790-85b2-3269762d6a55","Type":"ContainerStarted","Data":"25f6db2fc29fe084ee6d9b2bbd40a8d731059d718c61aa10f482f0f2e180fd52"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.950082 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" event={"ID":"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e","Type":"ContainerStarted","Data":"eecc3dd404e416162611c24accc03057e4a4f98a71e69ef19f667f15d61fac26"} Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.954344 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-955677c94-5qbwv" podStartSLOduration=4.294926819 podStartE2EDuration="49.954318486s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.216345751 +0000 UTC m=+1087.748859672" lastFinishedPulling="2025-11-28 11:40:55.875737418 +0000 UTC m=+1133.408251339" observedRunningTime="2025-11-28 11:40:57.950352342 +0000 UTC m=+1135.482866263" watchObservedRunningTime="2025-11-28 11:40:57.954318486 +0000 UTC m=+1135.486832407" Nov 28 11:40:57 crc kubenswrapper[4862]: I1128 11:40:57.968635 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" event={"ID":"9e68a5c6-829c-4897-b251-c181097105fc","Type":"ContainerStarted","Data":"bd32aa28ba8c20729cc92cd956311701963c6d2c08e48c229df3f98b797ebbfd"} Nov 28 11:40:58 crc kubenswrapper[4862]: E1128 11:40:58.094856 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" podUID="a836e80a-5070-4ac6-9d06-13f7dcb8f5a7" Nov 28 11:40:58 crc kubenswrapper[4862]: E1128 11:40:58.135525 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" podUID="d8d42319-792f-4aec-abcc-7158a1bf9f9c" Nov 28 11:40:58 crc kubenswrapper[4862]: E1128 11:40:58.135608 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" podUID="84002eaf-76bf-4a7e-a7ef-81c87d450270" Nov 28 11:40:58 crc kubenswrapper[4862]: I1128 11:40:58.977608 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" event={"ID":"4c8abe61-f6dd-4bf2-80e6-c13db3db907f","Type":"ContainerStarted","Data":"6d7c2b1db9a415c0aa991a6583f246e65c1529d6fd90cbc632499175c653ddc1"} Nov 28 11:40:58 crc kubenswrapper[4862]: I1128 11:40:58.977673 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" Nov 28 11:40:58 crc kubenswrapper[4862]: I1128 11:40:58.979856 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" event={"ID":"f4857ca4-a829-4205-b776-fa73faf12eac","Type":"ContainerStarted","Data":"ff24df8a0c7613fd6421aa4177bdb4062720bf9ed5ad296d1aa88b0223006958"} Nov 28 11:40:58 crc kubenswrapper[4862]: I1128 11:40:58.979990 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:40:58 crc kubenswrapper[4862]: I1128 11:40:58.982287 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" event={"ID":"36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e","Type":"ContainerStarted","Data":"b5e4d006f7992d7915694c59bd86cd33ff383a1c5da0d27ee0354513551ca9d5"} Nov 28 11:40:58 crc kubenswrapper[4862]: I1128 11:40:58.983368 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:40:58 crc kubenswrapper[4862]: E1128 11:40:58.984720 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" podUID="d8d42319-792f-4aec-abcc-7158a1bf9f9c" Nov 28 11:40:58 crc kubenswrapper[4862]: E1128 11:40:58.985363 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423\\\"\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" podUID="84002eaf-76bf-4a7e-a7ef-81c87d450270" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.001425 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" podStartSLOduration=5.966878472 podStartE2EDuration="51.001392715s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.684302064 +0000 UTC m=+1088.216815985" lastFinishedPulling="2025-11-28 11:40:55.718816307 +0000 UTC m=+1133.251330228" observedRunningTime="2025-11-28 11:40:58.998638296 +0000 UTC m=+1136.531152257" watchObservedRunningTime="2025-11-28 11:40:59.001392715 +0000 UTC m=+1136.533906656" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.026364 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" podStartSLOduration=42.296423546 podStartE2EDuration="51.026347294s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:47.207496838 +0000 UTC m=+1124.740010759" lastFinishedPulling="2025-11-28 11:40:55.937420586 +0000 UTC m=+1133.469934507" observedRunningTime="2025-11-28 11:40:59.018782116 +0000 UTC m=+1136.551296047" watchObservedRunningTime="2025-11-28 11:40:59.026347294 +0000 UTC m=+1136.558861215" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.043375 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.048158 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" podStartSLOduration=5.275360468 podStartE2EDuration="51.048140312s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.207009832 +0000 UTC m=+1087.739523753" lastFinishedPulling="2025-11-28 11:40:55.979789666 +0000 UTC m=+1133.512303597" observedRunningTime="2025-11-28 11:40:59.04354507 +0000 UTC m=+1136.576058991" watchObservedRunningTime="2025-11-28 11:40:59.048140312 +0000 UTC m=+1136.580654233" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.067557 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" podStartSLOduration=5.109197371 podStartE2EDuration="50.067540981s" podCreationTimestamp="2025-11-28 11:40:09 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.980008233 +0000 UTC m=+1088.512522154" lastFinishedPulling="2025-11-28 11:40:55.938351843 +0000 UTC m=+1133.470865764" observedRunningTime="2025-11-28 11:40:59.063536236 +0000 UTC m=+1136.596050157" watchObservedRunningTime="2025-11-28 11:40:59.067540981 +0000 UTC m=+1136.600054902" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.077338 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.080973 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.086567 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-6h5f6" podStartSLOduration=5.568906355 podStartE2EDuration="51.086527798s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.505285386 +0000 UTC m=+1088.037799307" lastFinishedPulling="2025-11-28 11:40:56.022906819 +0000 UTC m=+1133.555420750" observedRunningTime="2025-11-28 11:40:59.085076536 +0000 UTC m=+1136.617590457" watchObservedRunningTime="2025-11-28 11:40:59.086527798 +0000 UTC m=+1136.619041719" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.111550 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" podStartSLOduration=5.713453159 podStartE2EDuration="51.111521318s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.670715502 +0000 UTC m=+1088.203229423" lastFinishedPulling="2025-11-28 11:40:56.068783661 +0000 UTC m=+1133.601297582" observedRunningTime="2025-11-28 11:40:59.108748418 +0000 UTC m=+1136.641262339" watchObservedRunningTime="2025-11-28 11:40:59.111521318 +0000 UTC m=+1136.644035239" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.155959 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" podStartSLOduration=43.355144701 podStartE2EDuration="51.155940518s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:48.13805937 +0000 UTC m=+1125.670573331" lastFinishedPulling="2025-11-28 11:40:55.938855207 +0000 UTC m=+1133.471369148" observedRunningTime="2025-11-28 11:40:59.155143175 +0000 UTC m=+1136.687657096" watchObservedRunningTime="2025-11-28 11:40:59.155940518 +0000 UTC m=+1136.688454439" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.171890 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.205555 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" podStartSLOduration=5.981272347 podStartE2EDuration="51.205530687s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.793004926 +0000 UTC m=+1088.325518857" lastFinishedPulling="2025-11-28 11:40:56.017263276 +0000 UTC m=+1133.549777197" observedRunningTime="2025-11-28 11:40:59.200141021 +0000 UTC m=+1136.732654942" watchObservedRunningTime="2025-11-28 11:40:59.205530687 +0000 UTC m=+1136.738044608" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.218899 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" podStartSLOduration=5.275288305 podStartE2EDuration="51.218880841s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.082850075 +0000 UTC m=+1087.615363996" lastFinishedPulling="2025-11-28 11:40:56.026442611 +0000 UTC m=+1133.558956532" observedRunningTime="2025-11-28 11:40:59.215849004 +0000 UTC m=+1136.748362925" watchObservedRunningTime="2025-11-28 11:40:59.218880841 +0000 UTC m=+1136.751394752" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.242453 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" podStartSLOduration=7.469249189 podStartE2EDuration="51.24241778s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.505487352 +0000 UTC m=+1088.038001273" lastFinishedPulling="2025-11-28 11:40:54.278655943 +0000 UTC m=+1131.811169864" observedRunningTime="2025-11-28 11:40:59.232589706 +0000 UTC m=+1136.765103627" watchObservedRunningTime="2025-11-28 11:40:59.24241778 +0000 UTC m=+1136.774931701" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.251291 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.254410 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.292428 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tmq8w" podStartSLOduration=6.079046003 podStartE2EDuration="51.29240869s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.808326897 +0000 UTC m=+1088.340840818" lastFinishedPulling="2025-11-28 11:40:56.021689574 +0000 UTC m=+1133.554203505" observedRunningTime="2025-11-28 11:40:59.272358292 +0000 UTC m=+1136.804872213" watchObservedRunningTime="2025-11-28 11:40:59.29240869 +0000 UTC m=+1136.824922611" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.300485 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.302068 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-lnvbp" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.313544 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" podStartSLOduration=6.277061468 podStartE2EDuration="51.313511058s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.980742634 +0000 UTC m=+1088.513256555" lastFinishedPulling="2025-11-28 11:40:56.017192214 +0000 UTC m=+1133.549706145" observedRunningTime="2025-11-28 11:40:59.311597473 +0000 UTC m=+1136.844111394" watchObservedRunningTime="2025-11-28 11:40:59.313511058 +0000 UTC m=+1136.846024979" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.318470 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" podStartSLOduration=6.271902229 podStartE2EDuration="51.31845242s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.979808757 +0000 UTC m=+1088.512322678" lastFinishedPulling="2025-11-28 11:40:56.026358948 +0000 UTC m=+1133.558872869" observedRunningTime="2025-11-28 11:40:59.295991153 +0000 UTC m=+1136.828505074" watchObservedRunningTime="2025-11-28 11:40:59.31845242 +0000 UTC m=+1136.850966341" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.331697 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" podStartSLOduration=6.105182737 podStartE2EDuration="51.331668671s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.796501597 +0000 UTC m=+1088.329015518" lastFinishedPulling="2025-11-28 11:40:56.022987521 +0000 UTC m=+1133.555501452" observedRunningTime="2025-11-28 11:40:59.324804913 +0000 UTC m=+1136.857318834" watchObservedRunningTime="2025-11-28 11:40:59.331668671 +0000 UTC m=+1136.864182592" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.440252 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.442081 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-776s4" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.528205 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.530927 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xsqj2" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.603592 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.606739 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-54k27" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.652647 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.664829 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-jz7cq" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.894497 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" Nov 28 11:40:59 crc kubenswrapper[4862]: I1128 11:40:59.897552 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-p5n9j" Nov 28 11:41:01 crc kubenswrapper[4862]: I1128 11:41:01.004991 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" event={"ID":"c50b92c1-e947-4f41-abfe-278e1826a0c9","Type":"ContainerStarted","Data":"47c9735695e951f4ae4952b19030d6e6373e091e06cb19827c584a599bbf0e1b"} Nov 28 11:41:01 crc kubenswrapper[4862]: I1128 11:41:01.024161 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-dmqbn" podStartSLOduration=20.718571708 podStartE2EDuration="53.024137105s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.657282235 +0000 UTC m=+1088.189796156" lastFinishedPulling="2025-11-28 11:40:42.962847592 +0000 UTC m=+1120.495361553" observedRunningTime="2025-11-28 11:41:01.023860857 +0000 UTC m=+1138.556374848" watchObservedRunningTime="2025-11-28 11:41:01.024137105 +0000 UTC m=+1138.556651066" Nov 28 11:41:05 crc kubenswrapper[4862]: I1128 11:41:05.048841 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nxpb6" Nov 28 11:41:05 crc kubenswrapper[4862]: I1128 11:41:05.387949 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67glltc" Nov 28 11:41:05 crc kubenswrapper[4862]: E1128 11:41:05.578894 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" podUID="50d316aa-a1e4-4ca0-907a-52537f1b1218" Nov 28 11:41:06 crc kubenswrapper[4862]: I1128 11:41:06.044895 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" event={"ID":"a836e80a-5070-4ac6-9d06-13f7dcb8f5a7","Type":"ContainerStarted","Data":"089d0543160f8084b4b90f12bb175db67ceb26774edbc3035aaf31f5469a78ed"} Nov 28 11:41:06 crc kubenswrapper[4862]: I1128 11:41:06.046310 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" Nov 28 11:41:08 crc kubenswrapper[4862]: I1128 11:41:08.908348 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" Nov 28 11:41:08 crc kubenswrapper[4862]: I1128 11:41:08.911054 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-dj8zz" Nov 28 11:41:08 crc kubenswrapper[4862]: I1128 11:41:08.921395 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" Nov 28 11:41:08 crc kubenswrapper[4862]: I1128 11:41:08.923065 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-7vhdh" Nov 28 11:41:08 crc kubenswrapper[4862]: I1128 11:41:08.930003 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" podStartSLOduration=5.849845099 podStartE2EDuration="1m0.929964054s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.504634797 +0000 UTC m=+1088.037148718" lastFinishedPulling="2025-11-28 11:41:05.584753732 +0000 UTC m=+1143.117267673" observedRunningTime="2025-11-28 11:41:06.069132348 +0000 UTC m=+1143.601646269" watchObservedRunningTime="2025-11-28 11:41:08.929964054 +0000 UTC m=+1146.462478015" Nov 28 11:41:09 crc kubenswrapper[4862]: I1128 11:41:09.046056 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-v5t45" Nov 28 11:41:09 crc kubenswrapper[4862]: I1128 11:41:09.413836 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-vhg8w" Nov 28 11:41:12 crc kubenswrapper[4862]: I1128 11:41:12.103494 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" event={"ID":"a0567ace-33d0-4e27-a4c6-739b24a847c8","Type":"ContainerStarted","Data":"234ba79c11626e16c20e30a6810837346c3b640fe9965a1326bdba3299ef741f"} Nov 28 11:41:12 crc kubenswrapper[4862]: I1128 11:41:12.104611 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" Nov 28 11:41:12 crc kubenswrapper[4862]: I1128 11:41:12.132965 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" podStartSLOduration=3.8005304779999998 podStartE2EDuration="1m4.132937697s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.980617421 +0000 UTC m=+1088.513131332" lastFinishedPulling="2025-11-28 11:41:11.31302459 +0000 UTC m=+1148.845538551" observedRunningTime="2025-11-28 11:41:12.120975949 +0000 UTC m=+1149.653489900" watchObservedRunningTime="2025-11-28 11:41:12.132937697 +0000 UTC m=+1149.665451618" Nov 28 11:41:14 crc kubenswrapper[4862]: I1128 11:41:14.144555 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" event={"ID":"176a3ccc-ec57-4af5-9e09-cb958ed71a2b","Type":"ContainerStarted","Data":"d7c39a8cbe654066762732f8374b9b00d545cbee7a6b367979add6f7bb4e2410"} Nov 28 11:41:14 crc kubenswrapper[4862]: I1128 11:41:14.146447 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" Nov 28 11:41:14 crc kubenswrapper[4862]: I1128 11:41:14.183803 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" podStartSLOduration=2.8283313469999998 podStartE2EDuration="1m5.183775654s" podCreationTimestamp="2025-11-28 11:40:09 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.980837177 +0000 UTC m=+1088.513351098" lastFinishedPulling="2025-11-28 11:41:13.336281434 +0000 UTC m=+1150.868795405" observedRunningTime="2025-11-28 11:41:14.174606067 +0000 UTC m=+1151.707120018" watchObservedRunningTime="2025-11-28 11:41:14.183775654 +0000 UTC m=+1151.716289605" Nov 28 11:41:15 crc kubenswrapper[4862]: I1128 11:41:15.155570 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" event={"ID":"84002eaf-76bf-4a7e-a7ef-81c87d450270","Type":"ContainerStarted","Data":"3af97eb29279db48c1192b687a2d271824c0fe1d2e9c8fc0232b5ac321f49e12"} Nov 28 11:41:15 crc kubenswrapper[4862]: I1128 11:41:15.155861 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" Nov 28 11:41:15 crc kubenswrapper[4862]: I1128 11:41:15.158308 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" event={"ID":"d8d42319-792f-4aec-abcc-7158a1bf9f9c","Type":"ContainerStarted","Data":"f39a2a7ca81186b867ab687e8223b20fc30c0083c23409b51a2ae7d95399f0b6"} Nov 28 11:41:15 crc kubenswrapper[4862]: I1128 11:41:15.158680 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" Nov 28 11:41:15 crc kubenswrapper[4862]: I1128 11:41:15.179684 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" podStartSLOduration=3.797633498 podStartE2EDuration="1m7.179657698s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:11.004526219 +0000 UTC m=+1088.537040140" lastFinishedPulling="2025-11-28 11:41:14.386550379 +0000 UTC m=+1151.919064340" observedRunningTime="2025-11-28 11:41:15.172259103 +0000 UTC m=+1152.704773054" watchObservedRunningTime="2025-11-28 11:41:15.179657698 +0000 UTC m=+1152.712171639" Nov 28 11:41:15 crc kubenswrapper[4862]: I1128 11:41:15.199142 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" podStartSLOduration=3.7941160529999998 podStartE2EDuration="1m7.199118974s" podCreationTimestamp="2025-11-28 11:40:08 +0000 UTC" firstStartedPulling="2025-11-28 11:40:10.980444946 +0000 UTC m=+1088.512958867" lastFinishedPulling="2025-11-28 11:41:14.385447837 +0000 UTC m=+1151.917961788" observedRunningTime="2025-11-28 11:41:15.18901448 +0000 UTC m=+1152.721528411" watchObservedRunningTime="2025-11-28 11:41:15.199118974 +0000 UTC m=+1152.731632915" Nov 28 11:41:19 crc kubenswrapper[4862]: I1128 11:41:19.187786 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-7tqrc" Nov 28 11:41:19 crc kubenswrapper[4862]: I1128 11:41:19.503554 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-r6stc" Nov 28 11:41:19 crc kubenswrapper[4862]: I1128 11:41:19.711182 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-n2hrf" Nov 28 11:41:19 crc kubenswrapper[4862]: I1128 11:41:19.831365 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-trj7j" Nov 28 11:41:19 crc kubenswrapper[4862]: I1128 11:41:19.864776 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d4bgt" Nov 28 11:41:22 crc kubenswrapper[4862]: I1128 11:41:22.248933 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" event={"ID":"50d316aa-a1e4-4ca0-907a-52537f1b1218","Type":"ContainerStarted","Data":"ef773a4a5ecc0486ff9dd71dbf48c7193374749b5df13b3e6dd1e3af9eec7f5c"} Nov 28 11:41:22 crc kubenswrapper[4862]: I1128 11:41:22.272454 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtc9t" podStartSLOduration=2.925519192 podStartE2EDuration="1m13.272432014s" podCreationTimestamp="2025-11-28 11:40:09 +0000 UTC" firstStartedPulling="2025-11-28 11:40:11.059811962 +0000 UTC m=+1088.592325883" lastFinishedPulling="2025-11-28 11:41:21.406724774 +0000 UTC m=+1158.939238705" observedRunningTime="2025-11-28 11:41:22.264927186 +0000 UTC m=+1159.797441117" watchObservedRunningTime="2025-11-28 11:41:22.272432014 +0000 UTC m=+1159.804945935" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.856861 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-557f57d995-5ldvx"] Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.859169 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.872593 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-hwtpn" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.872617 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.872872 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.873569 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.884603 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-5ldvx"] Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.922340 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b55c19-a208-4195-b589-51d7e455963c-config\") pod \"dnsmasq-dns-557f57d995-5ldvx\" (UID: \"73b55c19-a208-4195-b589-51d7e455963c\") " pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.923081 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v24jd\" (UniqueName: \"kubernetes.io/projected/73b55c19-a208-4195-b589-51d7e455963c-kube-api-access-v24jd\") pod \"dnsmasq-dns-557f57d995-5ldvx\" (UID: \"73b55c19-a208-4195-b589-51d7e455963c\") " pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.935915 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-5kn9q"] Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.937925 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.942703 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 28 11:41:36 crc kubenswrapper[4862]: I1128 11:41:36.956631 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-5kn9q"] Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.024208 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzghz\" (UniqueName: \"kubernetes.io/projected/dd41eac3-532a-459f-9979-bdf55302c5af-kube-api-access-vzghz\") pod \"dnsmasq-dns-766fdc659c-5kn9q\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.024254 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-config\") pod \"dnsmasq-dns-766fdc659c-5kn9q\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.024339 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b55c19-a208-4195-b589-51d7e455963c-config\") pod \"dnsmasq-dns-557f57d995-5ldvx\" (UID: \"73b55c19-a208-4195-b589-51d7e455963c\") " pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.024376 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v24jd\" (UniqueName: \"kubernetes.io/projected/73b55c19-a208-4195-b589-51d7e455963c-kube-api-access-v24jd\") pod \"dnsmasq-dns-557f57d995-5ldvx\" (UID: \"73b55c19-a208-4195-b589-51d7e455963c\") " pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.024417 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-dns-svc\") pod \"dnsmasq-dns-766fdc659c-5kn9q\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.025610 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b55c19-a208-4195-b589-51d7e455963c-config\") pod \"dnsmasq-dns-557f57d995-5ldvx\" (UID: \"73b55c19-a208-4195-b589-51d7e455963c\") " pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.050836 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v24jd\" (UniqueName: \"kubernetes.io/projected/73b55c19-a208-4195-b589-51d7e455963c-kube-api-access-v24jd\") pod \"dnsmasq-dns-557f57d995-5ldvx\" (UID: \"73b55c19-a208-4195-b589-51d7e455963c\") " pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.124758 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-dns-svc\") pod \"dnsmasq-dns-766fdc659c-5kn9q\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.124824 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzghz\" (UniqueName: \"kubernetes.io/projected/dd41eac3-532a-459f-9979-bdf55302c5af-kube-api-access-vzghz\") pod \"dnsmasq-dns-766fdc659c-5kn9q\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.124845 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-config\") pod \"dnsmasq-dns-766fdc659c-5kn9q\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.125764 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-config\") pod \"dnsmasq-dns-766fdc659c-5kn9q\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.126150 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-dns-svc\") pod \"dnsmasq-dns-766fdc659c-5kn9q\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.151305 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzghz\" (UniqueName: \"kubernetes.io/projected/dd41eac3-532a-459f-9979-bdf55302c5af-kube-api-access-vzghz\") pod \"dnsmasq-dns-766fdc659c-5kn9q\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.176823 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.255762 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.660377 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-5ldvx"] Nov 28 11:41:37 crc kubenswrapper[4862]: I1128 11:41:37.746339 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-5kn9q"] Nov 28 11:41:37 crc kubenswrapper[4862]: W1128 11:41:37.752925 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd41eac3_532a_459f_9979_bdf55302c5af.slice/crio-b8474fab2afe9dd3a8c899bcec88c0a0521ca2316b01abe3580da357466695d3 WatchSource:0}: Error finding container b8474fab2afe9dd3a8c899bcec88c0a0521ca2316b01abe3580da357466695d3: Status 404 returned error can't find the container with id b8474fab2afe9dd3a8c899bcec88c0a0521ca2316b01abe3580da357466695d3 Nov 28 11:41:38 crc kubenswrapper[4862]: I1128 11:41:38.291729 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:41:38 crc kubenswrapper[4862]: I1128 11:41:38.291794 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:41:38 crc kubenswrapper[4862]: I1128 11:41:38.388697 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-557f57d995-5ldvx" event={"ID":"73b55c19-a208-4195-b589-51d7e455963c","Type":"ContainerStarted","Data":"54074ba66bf24b4320591fa9c68a1316825f0a25d5815d41380f589d8207256e"} Nov 28 11:41:38 crc kubenswrapper[4862]: I1128 11:41:38.390754 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" event={"ID":"dd41eac3-532a-459f-9979-bdf55302c5af","Type":"ContainerStarted","Data":"b8474fab2afe9dd3a8c899bcec88c0a0521ca2316b01abe3580da357466695d3"} Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.616347 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-5ldvx"] Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.638634 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-qrkhk"] Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.639939 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.656899 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-qrkhk"] Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.663944 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5tdx\" (UniqueName: \"kubernetes.io/projected/2c23915a-f7a4-4334-aa70-31bf066f10d3-kube-api-access-k5tdx\") pod \"dnsmasq-dns-57dc4c6697-qrkhk\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.664059 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-dns-svc\") pod \"dnsmasq-dns-57dc4c6697-qrkhk\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.664140 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-config\") pod \"dnsmasq-dns-57dc4c6697-qrkhk\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.765245 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-dns-svc\") pod \"dnsmasq-dns-57dc4c6697-qrkhk\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.765315 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-config\") pod \"dnsmasq-dns-57dc4c6697-qrkhk\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.765347 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5tdx\" (UniqueName: \"kubernetes.io/projected/2c23915a-f7a4-4334-aa70-31bf066f10d3-kube-api-access-k5tdx\") pod \"dnsmasq-dns-57dc4c6697-qrkhk\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.766988 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-config\") pod \"dnsmasq-dns-57dc4c6697-qrkhk\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.769422 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-dns-svc\") pod \"dnsmasq-dns-57dc4c6697-qrkhk\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.787643 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5tdx\" (UniqueName: \"kubernetes.io/projected/2c23915a-f7a4-4334-aa70-31bf066f10d3-kube-api-access-k5tdx\") pod \"dnsmasq-dns-57dc4c6697-qrkhk\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.955166 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-5kn9q"] Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.958909 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.977706 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-c6fvq"] Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.980804 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:39 crc kubenswrapper[4862]: I1128 11:41:39.993416 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-c6fvq"] Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.078515 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-config\") pod \"dnsmasq-dns-8446fd7c75-c6fvq\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.078578 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-dns-svc\") pod \"dnsmasq-dns-8446fd7c75-c6fvq\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.078650 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4w82\" (UniqueName: \"kubernetes.io/projected/fa0af7e6-4966-4915-9807-d2055f09c111-kube-api-access-m4w82\") pod \"dnsmasq-dns-8446fd7c75-c6fvq\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.180657 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-config\") pod \"dnsmasq-dns-8446fd7c75-c6fvq\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.180701 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-dns-svc\") pod \"dnsmasq-dns-8446fd7c75-c6fvq\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.180766 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4w82\" (UniqueName: \"kubernetes.io/projected/fa0af7e6-4966-4915-9807-d2055f09c111-kube-api-access-m4w82\") pod \"dnsmasq-dns-8446fd7c75-c6fvq\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.181780 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-dns-svc\") pod \"dnsmasq-dns-8446fd7c75-c6fvq\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.184836 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-config\") pod \"dnsmasq-dns-8446fd7c75-c6fvq\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.212134 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4w82\" (UniqueName: \"kubernetes.io/projected/fa0af7e6-4966-4915-9807-d2055f09c111-kube-api-access-m4w82\") pod \"dnsmasq-dns-8446fd7c75-c6fvq\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.314550 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.561511 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-qrkhk"] Nov 28 11:41:40 crc kubenswrapper[4862]: W1128 11:41:40.569037 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c23915a_f7a4_4334_aa70_31bf066f10d3.slice/crio-785c4f7932185b0be0aae34dbc7ced0a30b727b3c9c8665eb0cac2406c1d7046 WatchSource:0}: Error finding container 785c4f7932185b0be0aae34dbc7ced0a30b727b3c9c8665eb0cac2406c1d7046: Status 404 returned error can't find the container with id 785c4f7932185b0be0aae34dbc7ced0a30b727b3c9c8665eb0cac2406c1d7046 Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.790441 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-c6fvq"] Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.832033 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.833374 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.835514 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.836538 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.836600 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.836673 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-b8qqx" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.836847 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.836862 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.836898 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.864058 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.895660 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.895720 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.895758 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.895826 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.895991 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-pod-info\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.896151 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.896188 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-server-conf\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.896209 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.896297 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.896319 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.896382 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgkwp\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-kube-api-access-rgkwp\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997291 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997344 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997386 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997405 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997423 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-pod-info\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997446 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997460 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-server-conf\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997474 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997511 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997525 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.997553 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgkwp\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-kube-api-access-rgkwp\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.998235 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.998349 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.998436 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.998473 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.999209 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:40 crc kubenswrapper[4862]: I1128 11:41:40.999960 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-server-conf\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.005403 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.005562 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-pod-info\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.007741 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.014811 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.015247 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgkwp\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-kube-api-access-rgkwp\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.021449 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " pod="openstack/rabbitmq-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.100693 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.105590 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.111726 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.118640 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.118813 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.118937 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hzbmn" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.119051 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.119703 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.119821 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.127573 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.167784 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202605 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202647 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202672 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfwrl\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-kube-api-access-wfwrl\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202696 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/527b2bce-1186-4186-8992-a3a63d2d2b22-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202712 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202732 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202787 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202800 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202823 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/527b2bce-1186-4186-8992-a3a63d2d2b22-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202840 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.202857 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.303687 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/527b2bce-1186-4186-8992-a3a63d2d2b22-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.303754 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.303803 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.303936 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.303964 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.303989 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfwrl\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-kube-api-access-wfwrl\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.304098 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.304120 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/527b2bce-1186-4186-8992-a3a63d2d2b22-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.304146 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.304250 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.304280 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.305636 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.305745 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.307763 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.308023 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.308551 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.315676 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/527b2bce-1186-4186-8992-a3a63d2d2b22-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.317025 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.319364 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.322703 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/527b2bce-1186-4186-8992-a3a63d2d2b22-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.334543 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.336511 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfwrl\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-kube-api-access-wfwrl\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.343811 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.424223 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" event={"ID":"2c23915a-f7a4-4334-aa70-31bf066f10d3","Type":"ContainerStarted","Data":"785c4f7932185b0be0aae34dbc7ced0a30b727b3c9c8665eb0cac2406c1d7046"} Nov 28 11:41:41 crc kubenswrapper[4862]: I1128 11:41:41.437955 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.139764 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.142331 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.144840 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-jvlkf" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.145498 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.145749 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.147133 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.155081 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.162591 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.320137 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-default\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.320204 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.320229 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.320343 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.320420 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-kolla-config\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.320478 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.320628 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.320703 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44qrp\" (UniqueName: \"kubernetes.io/projected/f01191bd-12ff-43ea-9788-381a997a3c08-kube-api-access-44qrp\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.422979 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-default\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.423051 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.423076 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.423142 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.423170 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-kolla-config\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.423192 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.423209 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.423226 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44qrp\" (UniqueName: \"kubernetes.io/projected/f01191bd-12ff-43ea-9788-381a997a3c08-kube-api-access-44qrp\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.423307 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.423916 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-default\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.424200 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.428153 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.450009 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44qrp\" (UniqueName: \"kubernetes.io/projected/f01191bd-12ff-43ea-9788-381a997a3c08-kube-api-access-44qrp\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.454558 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.454925 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-kolla-config\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.455680 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.462958 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " pod="openstack/openstack-galera-0" Nov 28 11:41:42 crc kubenswrapper[4862]: I1128 11:41:42.477411 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.696382 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.698026 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.703381 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-tptnf" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.725689 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.730048 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.730262 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.742063 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.851019 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.851075 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.851189 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jsjq\" (UniqueName: \"kubernetes.io/projected/27c60127-ee39-4d28-89f7-e5225ddc63d7-kube-api-access-7jsjq\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.851262 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.851289 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.851505 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.851539 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.851635 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.908473 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.909644 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.914454 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.916809 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-z8bxm" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.917345 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.920701 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.953154 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.953230 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.953280 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.953312 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jsjq\" (UniqueName: \"kubernetes.io/projected/27c60127-ee39-4d28-89f7-e5225ddc63d7-kube-api-access-7jsjq\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.953345 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.953366 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.953462 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.953491 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.953862 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.954074 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.954148 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.955498 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.956221 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.962871 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.974635 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.975526 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jsjq\" (UniqueName: \"kubernetes.io/projected/27c60127-ee39-4d28-89f7-e5225ddc63d7-kube-api-access-7jsjq\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:43 crc kubenswrapper[4862]: I1128 11:41:43.983509 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.015281 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.064802 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcjnt\" (UniqueName: \"kubernetes.io/projected/4559a88d-6d07-44d0-a829-c35fd773fcfb-kube-api-access-hcjnt\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.064877 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.065124 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-config-data\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.065228 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-kolla-config\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.065302 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.166594 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-config-data\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.166664 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-kolla-config\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.166694 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.166725 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcjnt\" (UniqueName: \"kubernetes.io/projected/4559a88d-6d07-44d0-a829-c35fd773fcfb-kube-api-access-hcjnt\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.166742 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.167773 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-config-data\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.168166 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-kolla-config\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.170231 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.170913 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.181547 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcjnt\" (UniqueName: \"kubernetes.io/projected/4559a88d-6d07-44d0-a829-c35fd773fcfb-kube-api-access-hcjnt\") pod \"memcached-0\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " pod="openstack/memcached-0" Nov 28 11:41:44 crc kubenswrapper[4862]: I1128 11:41:44.226084 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 11:41:45 crc kubenswrapper[4862]: W1128 11:41:45.354905 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa0af7e6_4966_4915_9807_d2055f09c111.slice/crio-2bb8690f3555c12056e09baca4783a483520c623094d98344166f0f1747ee4d7 WatchSource:0}: Error finding container 2bb8690f3555c12056e09baca4783a483520c623094d98344166f0f1747ee4d7: Status 404 returned error can't find the container with id 2bb8690f3555c12056e09baca4783a483520c623094d98344166f0f1747ee4d7 Nov 28 11:41:45 crc kubenswrapper[4862]: I1128 11:41:45.479328 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" event={"ID":"fa0af7e6-4966-4915-9807-d2055f09c111","Type":"ContainerStarted","Data":"2bb8690f3555c12056e09baca4783a483520c623094d98344166f0f1747ee4d7"} Nov 28 11:41:46 crc kubenswrapper[4862]: I1128 11:41:46.138300 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:41:46 crc kubenswrapper[4862]: I1128 11:41:46.139652 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 11:41:46 crc kubenswrapper[4862]: I1128 11:41:46.144396 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-77z8f" Nov 28 11:41:46 crc kubenswrapper[4862]: I1128 11:41:46.151081 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:41:46 crc kubenswrapper[4862]: I1128 11:41:46.304479 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ns99\" (UniqueName: \"kubernetes.io/projected/264bc03e-9ff0-44ce-b7d3-05111d5b6483-kube-api-access-4ns99\") pod \"kube-state-metrics-0\" (UID: \"264bc03e-9ff0-44ce-b7d3-05111d5b6483\") " pod="openstack/kube-state-metrics-0" Nov 28 11:41:46 crc kubenswrapper[4862]: I1128 11:41:46.406303 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ns99\" (UniqueName: \"kubernetes.io/projected/264bc03e-9ff0-44ce-b7d3-05111d5b6483-kube-api-access-4ns99\") pod \"kube-state-metrics-0\" (UID: \"264bc03e-9ff0-44ce-b7d3-05111d5b6483\") " pod="openstack/kube-state-metrics-0" Nov 28 11:41:46 crc kubenswrapper[4862]: I1128 11:41:46.424824 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ns99\" (UniqueName: \"kubernetes.io/projected/264bc03e-9ff0-44ce-b7d3-05111d5b6483-kube-api-access-4ns99\") pod \"kube-state-metrics-0\" (UID: \"264bc03e-9ff0-44ce-b7d3-05111d5b6483\") " pod="openstack/kube-state-metrics-0" Nov 28 11:41:46 crc kubenswrapper[4862]: I1128 11:41:46.461700 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.578338 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-hdvv6"] Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.581391 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.584141 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-fgbsl" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.585062 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.585235 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hkgwr"] Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.586449 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.589705 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.593019 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-hdvv6"] Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.604163 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hkgwr"] Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.663879 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-scripts\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.663939 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtpb9\" (UniqueName: \"kubernetes.io/projected/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-kube-api-access-jtpb9\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664009 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-log\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664041 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-log-ovn\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664068 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run-ovn\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664121 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1bdf176-776a-4211-8b43-ff90225de25a-scripts\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664184 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-etc-ovs\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664263 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-run\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664296 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-lib\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664328 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-combined-ca-bundle\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664389 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-ovn-controller-tls-certs\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664426 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.664441 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8skk\" (UniqueName: \"kubernetes.io/projected/b1bdf176-776a-4211-8b43-ff90225de25a-kube-api-access-t8skk\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766409 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-lib\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766495 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-combined-ca-bundle\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766563 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-ovn-controller-tls-certs\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766646 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766678 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8skk\" (UniqueName: \"kubernetes.io/projected/b1bdf176-776a-4211-8b43-ff90225de25a-kube-api-access-t8skk\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766715 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-scripts\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766745 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtpb9\" (UniqueName: \"kubernetes.io/projected/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-kube-api-access-jtpb9\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766793 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-log\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766828 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-log-ovn\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766879 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run-ovn\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766911 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1bdf176-776a-4211-8b43-ff90225de25a-scripts\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.766965 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-etc-ovs\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.767080 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-run\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.767783 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-log\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.767884 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.767976 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-run\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.767983 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-lib\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.767987 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run-ovn\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.768057 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-etc-ovs\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.768179 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-log-ovn\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.770810 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1bdf176-776a-4211-8b43-ff90225de25a-scripts\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.771780 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-scripts\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.773437 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-ovn-controller-tls-certs\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.781739 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-combined-ca-bundle\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.796879 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8skk\" (UniqueName: \"kubernetes.io/projected/b1bdf176-776a-4211-8b43-ff90225de25a-kube-api-access-t8skk\") pod \"ovn-controller-hkgwr\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.814931 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtpb9\" (UniqueName: \"kubernetes.io/projected/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-kube-api-access-jtpb9\") pod \"ovn-controller-ovs-hdvv6\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.913494 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:41:49 crc kubenswrapper[4862]: I1128 11:41:49.927306 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hkgwr" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.427957 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.430606 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.444388 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.446586 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.449662 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-lvvms" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.451686 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.452991 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.473396 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.526425 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.526510 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.526543 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69897166-b5c9-404e-ac62-0b55f0960278-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.526563 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg8xp\" (UniqueName: \"kubernetes.io/projected/69897166-b5c9-404e-ac62-0b55f0960278-kube-api-access-mg8xp\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.526669 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.526699 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.526748 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-config\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.526859 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.627843 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-config\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.627943 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.628040 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.628127 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.628170 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69897166-b5c9-404e-ac62-0b55f0960278-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.628200 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg8xp\" (UniqueName: \"kubernetes.io/projected/69897166-b5c9-404e-ac62-0b55f0960278-kube-api-access-mg8xp\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.628550 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.628954 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69897166-b5c9-404e-ac62-0b55f0960278-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.629200 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.629272 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.629280 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-config\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.630437 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.634240 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.634514 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.637037 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.648449 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg8xp\" (UniqueName: \"kubernetes.io/projected/69897166-b5c9-404e-ac62-0b55f0960278-kube-api-access-mg8xp\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.671308 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:52 crc kubenswrapper[4862]: I1128 11:41:52.768732 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.177169 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.178778 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.181841 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.182833 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-5l9mb" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.183139 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.183921 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.191575 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.244841 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cr4v\" (UniqueName: \"kubernetes.io/projected/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-kube-api-access-8cr4v\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.244967 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.245010 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.245083 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.245158 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.245225 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.245282 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-config\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.245332 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.347459 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cr4v\" (UniqueName: \"kubernetes.io/projected/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-kube-api-access-8cr4v\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.347826 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.347867 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.347916 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.347948 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.347988 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.348027 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-config\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.348072 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.349049 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.350082 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.351455 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-config\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.352595 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.357801 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.367614 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.368774 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.374763 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cr4v\" (UniqueName: \"kubernetes.io/projected/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-kube-api-access-8cr4v\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.394519 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " pod="openstack/ovsdbserver-nb-0" Nov 28 11:41:53 crc kubenswrapper[4862]: I1128 11:41:53.514528 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 11:42:00 crc kubenswrapper[4862]: I1128 11:42:00.172280 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 11:42:00 crc kubenswrapper[4862]: I1128 11:42:00.442973 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 11:42:00 crc kubenswrapper[4862]: E1128 11:42:00.769566 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Nov 28 11:42:00 crc kubenswrapper[4862]: E1128 11:42:00.769767 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v24jd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-557f57d995-5ldvx_openstack(73b55c19-a208-4195-b589-51d7e455963c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:42:00 crc kubenswrapper[4862]: E1128 11:42:00.771186 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-557f57d995-5ldvx" podUID="73b55c19-a208-4195-b589-51d7e455963c" Nov 28 11:42:00 crc kubenswrapper[4862]: W1128 11:42:00.778152 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod527b2bce_1186_4186_8992_a3a63d2d2b22.slice/crio-bbf72032ab897813bfd4f4fed577a12e1dbbf694c3bfa01700bfc71a8e549018 WatchSource:0}: Error finding container bbf72032ab897813bfd4f4fed577a12e1dbbf694c3bfa01700bfc71a8e549018: Status 404 returned error can't find the container with id bbf72032ab897813bfd4f4fed577a12e1dbbf694c3bfa01700bfc71a8e549018 Nov 28 11:42:00 crc kubenswrapper[4862]: E1128 11:42:00.786544 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Nov 28 11:42:00 crc kubenswrapper[4862]: E1128 11:42:00.786721 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k5tdx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57dc4c6697-qrkhk_openstack(2c23915a-f7a4-4334-aa70-31bf066f10d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:42:00 crc kubenswrapper[4862]: E1128 11:42:00.787875 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" podUID="2c23915a-f7a4-4334-aa70-31bf066f10d3" Nov 28 11:42:00 crc kubenswrapper[4862]: E1128 11:42:00.814272 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Nov 28 11:42:00 crc kubenswrapper[4862]: E1128 11:42:00.814489 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vzghz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-766fdc659c-5kn9q_openstack(dd41eac3-532a-459f-9979-bdf55302c5af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:42:00 crc kubenswrapper[4862]: E1128 11:42:00.815650 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" podUID="dd41eac3-532a-459f-9979-bdf55302c5af" Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.068915 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 11:42:01 crc kubenswrapper[4862]: W1128 11:42:01.117400 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88b7fd5f_d4cc_4bd9_b9ec_039cf4a25356.slice/crio-c4724491b275a3f9f3046ef2356616d2d16720ef4c64f64f138d4a6d71865a54 WatchSource:0}: Error finding container c4724491b275a3f9f3046ef2356616d2d16720ef4c64f64f138d4a6d71865a54: Status 404 returned error can't find the container with id c4724491b275a3f9f3046ef2356616d2d16720ef4c64f64f138d4a6d71865a54 Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.486688 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 28 11:42:01 crc kubenswrapper[4862]: W1128 11:42:01.495758 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4559a88d_6d07_44d0_a829_c35fd773fcfb.slice/crio-9ee7bf639c6fcd4781e3e20eed08767a0f373457bcd9bab0bb0a4c93cd017667 WatchSource:0}: Error finding container 9ee7bf639c6fcd4781e3e20eed08767a0f373457bcd9bab0bb0a4c93cd017667: Status 404 returned error can't find the container with id 9ee7bf639c6fcd4781e3e20eed08767a0f373457bcd9bab0bb0a4c93cd017667 Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.500249 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:42:01 crc kubenswrapper[4862]: W1128 11:42:01.503190 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod264bc03e_9ff0_44ce_b7d3_05111d5b6483.slice/crio-b3246f0ffa7717c844ee023f89483fdbb0fe8058f206dcf92d6188b0fc6d998c WatchSource:0}: Error finding container b3246f0ffa7717c844ee023f89483fdbb0fe8058f206dcf92d6188b0fc6d998c: Status 404 returned error can't find the container with id b3246f0ffa7717c844ee023f89483fdbb0fe8058f206dcf92d6188b0fc6d998c Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.592060 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-hdvv6"] Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.614472 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.623152 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hkgwr"] Nov 28 11:42:01 crc kubenswrapper[4862]: W1128 11:42:01.624711 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1bdf176_776a_4211_8b43_ff90225de25a.slice/crio-ff2430e6a4891c2e18501cf0209ccc7e522f0465b6639bd7b47c8dde544ce99a WatchSource:0}: Error finding container ff2430e6a4891c2e18501cf0209ccc7e522f0465b6639bd7b47c8dde544ce99a: Status 404 returned error can't find the container with id ff2430e6a4891c2e18501cf0209ccc7e522f0465b6639bd7b47c8dde544ce99a Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.643213 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hkgwr" event={"ID":"b1bdf176-776a-4211-8b43-ff90225de25a","Type":"ContainerStarted","Data":"ff2430e6a4891c2e18501cf0209ccc7e522f0465b6639bd7b47c8dde544ce99a"} Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.645370 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"527b2bce-1186-4186-8992-a3a63d2d2b22","Type":"ContainerStarted","Data":"bbf72032ab897813bfd4f4fed577a12e1dbbf694c3bfa01700bfc71a8e549018"} Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.648979 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hdvv6" event={"ID":"ba7cc6b9-1443-45de-84f4-4ec3b92e699c","Type":"ContainerStarted","Data":"46f4a765ad8066346562b5840ee471cf98f1740cbd169e69b17951f103ae325e"} Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.652444 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"264bc03e-9ff0-44ce-b7d3-05111d5b6483","Type":"ContainerStarted","Data":"b3246f0ffa7717c844ee023f89483fdbb0fe8058f206dcf92d6188b0fc6d998c"} Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.654142 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"27c60127-ee39-4d28-89f7-e5225ddc63d7","Type":"ContainerStarted","Data":"df7b749c6694f2a235cbb438e4169480bec39c6cf7788018bd36c3c069c46d73"} Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.655827 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356","Type":"ContainerStarted","Data":"c4724491b275a3f9f3046ef2356616d2d16720ef4c64f64f138d4a6d71865a54"} Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.658641 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f01191bd-12ff-43ea-9788-381a997a3c08","Type":"ContainerStarted","Data":"4123c28387f62633cc6d3b318c5591f34816fcef70e767fa769f385b325aedc3"} Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.660717 4862 generic.go:334] "Generic (PLEG): container finished" podID="fa0af7e6-4966-4915-9807-d2055f09c111" containerID="76c619f1dbd52eb0e8a5b8c88d7dd5661d3fdec55812c8feb8f4fca1b5a9d61f" exitCode=0 Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.660832 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" event={"ID":"fa0af7e6-4966-4915-9807-d2055f09c111","Type":"ContainerDied","Data":"76c619f1dbd52eb0e8a5b8c88d7dd5661d3fdec55812c8feb8f4fca1b5a9d61f"} Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.662059 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4559a88d-6d07-44d0-a829-c35fd773fcfb","Type":"ContainerStarted","Data":"9ee7bf639c6fcd4781e3e20eed08767a0f373457bcd9bab0bb0a4c93cd017667"} Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.680796 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 11:42:01 crc kubenswrapper[4862]: I1128 11:42:01.809962 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 11:42:01 crc kubenswrapper[4862]: E1128 11:42:01.888925 4862 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 28 11:42:01 crc kubenswrapper[4862]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/2c23915a-f7a4-4334-aa70-31bf066f10d3/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 28 11:42:01 crc kubenswrapper[4862]: > podSandboxID="785c4f7932185b0be0aae34dbc7ced0a30b727b3c9c8665eb0cac2406c1d7046" Nov 28 11:42:01 crc kubenswrapper[4862]: E1128 11:42:01.889080 4862 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 28 11:42:01 crc kubenswrapper[4862]: init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k5tdx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57dc4c6697-qrkhk_openstack(2c23915a-f7a4-4334-aa70-31bf066f10d3): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/2c23915a-f7a4-4334-aa70-31bf066f10d3/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 28 11:42:01 crc kubenswrapper[4862]: > logger="UnhandledError" Nov 28 11:42:01 crc kubenswrapper[4862]: E1128 11:42:01.890362 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/2c23915a-f7a4-4334-aa70-31bf066f10d3/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" podUID="2c23915a-f7a4-4334-aa70-31bf066f10d3" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.061329 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.108796 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.236599 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-dns-svc\") pod \"dd41eac3-532a-459f-9979-bdf55302c5af\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.236658 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzghz\" (UniqueName: \"kubernetes.io/projected/dd41eac3-532a-459f-9979-bdf55302c5af-kube-api-access-vzghz\") pod \"dd41eac3-532a-459f-9979-bdf55302c5af\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.236691 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v24jd\" (UniqueName: \"kubernetes.io/projected/73b55c19-a208-4195-b589-51d7e455963c-kube-api-access-v24jd\") pod \"73b55c19-a208-4195-b589-51d7e455963c\" (UID: \"73b55c19-a208-4195-b589-51d7e455963c\") " Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.236740 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b55c19-a208-4195-b589-51d7e455963c-config\") pod \"73b55c19-a208-4195-b589-51d7e455963c\" (UID: \"73b55c19-a208-4195-b589-51d7e455963c\") " Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.236769 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-config\") pod \"dd41eac3-532a-459f-9979-bdf55302c5af\" (UID: \"dd41eac3-532a-459f-9979-bdf55302c5af\") " Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.237908 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-config" (OuterVolumeSpecName: "config") pod "dd41eac3-532a-459f-9979-bdf55302c5af" (UID: "dd41eac3-532a-459f-9979-bdf55302c5af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.239264 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd41eac3-532a-459f-9979-bdf55302c5af" (UID: "dd41eac3-532a-459f-9979-bdf55302c5af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.239579 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b55c19-a208-4195-b589-51d7e455963c-config" (OuterVolumeSpecName: "config") pod "73b55c19-a208-4195-b589-51d7e455963c" (UID: "73b55c19-a208-4195-b589-51d7e455963c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.241620 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd41eac3-532a-459f-9979-bdf55302c5af-kube-api-access-vzghz" (OuterVolumeSpecName: "kube-api-access-vzghz") pod "dd41eac3-532a-459f-9979-bdf55302c5af" (UID: "dd41eac3-532a-459f-9979-bdf55302c5af"). InnerVolumeSpecName "kube-api-access-vzghz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.241703 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b55c19-a208-4195-b589-51d7e455963c-kube-api-access-v24jd" (OuterVolumeSpecName: "kube-api-access-v24jd") pod "73b55c19-a208-4195-b589-51d7e455963c" (UID: "73b55c19-a208-4195-b589-51d7e455963c"). InnerVolumeSpecName "kube-api-access-v24jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.338612 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzghz\" (UniqueName: \"kubernetes.io/projected/dd41eac3-532a-459f-9979-bdf55302c5af-kube-api-access-vzghz\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.338644 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v24jd\" (UniqueName: \"kubernetes.io/projected/73b55c19-a208-4195-b589-51d7e455963c-kube-api-access-v24jd\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.338655 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b55c19-a208-4195-b589-51d7e455963c-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.338664 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.338962 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd41eac3-532a-459f-9979-bdf55302c5af-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.673568 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-557f57d995-5ldvx" event={"ID":"73b55c19-a208-4195-b589-51d7e455963c","Type":"ContainerDied","Data":"54074ba66bf24b4320591fa9c68a1316825f0a25d5815d41380f589d8207256e"} Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.673664 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-5ldvx" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.675952 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b","Type":"ContainerStarted","Data":"d9f88fdac33b734215e25545d9ffc70700d4da39e0489ff0ff9f805add4a1836"} Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.677279 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69897166-b5c9-404e-ac62-0b55f0960278","Type":"ContainerStarted","Data":"a3cfa1403df510c3ae69baccf10926138e6a55c6506dc9d3b5be5956bb888c94"} Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.679131 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" event={"ID":"fa0af7e6-4966-4915-9807-d2055f09c111","Type":"ContainerStarted","Data":"ee0d80ae8fee71d34c192ca3acdb1458773fb691afab6adc3870c00b988311d1"} Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.679247 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.680594 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" event={"ID":"dd41eac3-532a-459f-9979-bdf55302c5af","Type":"ContainerDied","Data":"b8474fab2afe9dd3a8c899bcec88c0a0521ca2316b01abe3580da357466695d3"} Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.680762 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-5kn9q" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.700246 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" podStartSLOduration=8.151528098 podStartE2EDuration="23.700229459s" podCreationTimestamp="2025-11-28 11:41:39 +0000 UTC" firstStartedPulling="2025-11-28 11:41:45.360352042 +0000 UTC m=+1182.892866023" lastFinishedPulling="2025-11-28 11:42:00.909053463 +0000 UTC m=+1198.441567384" observedRunningTime="2025-11-28 11:42:02.696730386 +0000 UTC m=+1200.229244307" watchObservedRunningTime="2025-11-28 11:42:02.700229459 +0000 UTC m=+1200.232743380" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.738855 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-5ldvx"] Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.744632 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-5ldvx"] Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.771736 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-5kn9q"] Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.777741 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-5kn9q"] Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.856940 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b55c19-a208-4195-b589-51d7e455963c" path="/var/lib/kubelet/pods/73b55c19-a208-4195-b589-51d7e455963c/volumes" Nov 28 11:42:02 crc kubenswrapper[4862]: I1128 11:42:02.857311 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd41eac3-532a-459f-9979-bdf55302c5af" path="/var/lib/kubelet/pods/dd41eac3-532a-459f-9979-bdf55302c5af/volumes" Nov 28 11:42:08 crc kubenswrapper[4862]: I1128 11:42:08.292366 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:42:08 crc kubenswrapper[4862]: I1128 11:42:08.293235 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:42:10 crc kubenswrapper[4862]: I1128 11:42:10.318476 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:42:10 crc kubenswrapper[4862]: I1128 11:42:10.376590 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-qrkhk"] Nov 28 11:42:10 crc kubenswrapper[4862]: I1128 11:42:10.758672 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"27c60127-ee39-4d28-89f7-e5225ddc63d7","Type":"ContainerStarted","Data":"fcd0cc7dfd92fb5cac96b835cfe5b465117571bc62e3995b22db01397ac53877"} Nov 28 11:42:11 crc kubenswrapper[4862]: I1128 11:42:11.769861 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356","Type":"ContainerStarted","Data":"28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245"} Nov 28 11:42:11 crc kubenswrapper[4862]: I1128 11:42:11.774927 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" event={"ID":"2c23915a-f7a4-4334-aa70-31bf066f10d3","Type":"ContainerDied","Data":"785c4f7932185b0be0aae34dbc7ced0a30b727b3c9c8665eb0cac2406c1d7046"} Nov 28 11:42:11 crc kubenswrapper[4862]: I1128 11:42:11.775004 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="785c4f7932185b0be0aae34dbc7ced0a30b727b3c9c8665eb0cac2406c1d7046" Nov 28 11:42:11 crc kubenswrapper[4862]: I1128 11:42:11.777758 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:42:11 crc kubenswrapper[4862]: I1128 11:42:11.909955 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5tdx\" (UniqueName: \"kubernetes.io/projected/2c23915a-f7a4-4334-aa70-31bf066f10d3-kube-api-access-k5tdx\") pod \"2c23915a-f7a4-4334-aa70-31bf066f10d3\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " Nov 28 11:42:11 crc kubenswrapper[4862]: I1128 11:42:11.910027 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-config\") pod \"2c23915a-f7a4-4334-aa70-31bf066f10d3\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " Nov 28 11:42:11 crc kubenswrapper[4862]: I1128 11:42:11.910084 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-dns-svc\") pod \"2c23915a-f7a4-4334-aa70-31bf066f10d3\" (UID: \"2c23915a-f7a4-4334-aa70-31bf066f10d3\") " Nov 28 11:42:11 crc kubenswrapper[4862]: I1128 11:42:11.954373 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c23915a-f7a4-4334-aa70-31bf066f10d3-kube-api-access-k5tdx" (OuterVolumeSpecName: "kube-api-access-k5tdx") pod "2c23915a-f7a4-4334-aa70-31bf066f10d3" (UID: "2c23915a-f7a4-4334-aa70-31bf066f10d3"). InnerVolumeSpecName "kube-api-access-k5tdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.011719 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5tdx\" (UniqueName: \"kubernetes.io/projected/2c23915a-f7a4-4334-aa70-31bf066f10d3-kube-api-access-k5tdx\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.127133 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-config" (OuterVolumeSpecName: "config") pod "2c23915a-f7a4-4334-aa70-31bf066f10d3" (UID: "2c23915a-f7a4-4334-aa70-31bf066f10d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.215585 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.252756 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2c23915a-f7a4-4334-aa70-31bf066f10d3" (UID: "2c23915a-f7a4-4334-aa70-31bf066f10d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.317556 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c23915a-f7a4-4334-aa70-31bf066f10d3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.782751 4862 generic.go:334] "Generic (PLEG): container finished" podID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerID="b26a7efa52fe59cc26535571cd7c3b5c387aba3ffa9eb88168ffcfe54e853cf0" exitCode=0 Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.782919 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hdvv6" event={"ID":"ba7cc6b9-1443-45de-84f4-4ec3b92e699c","Type":"ContainerDied","Data":"b26a7efa52fe59cc26535571cd7c3b5c387aba3ffa9eb88168ffcfe54e853cf0"} Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.786026 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f01191bd-12ff-43ea-9788-381a997a3c08","Type":"ContainerStarted","Data":"7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755"} Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.788330 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"264bc03e-9ff0-44ce-b7d3-05111d5b6483","Type":"ContainerStarted","Data":"c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11"} Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.788939 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.793439 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b","Type":"ContainerStarted","Data":"ee8d7e9a14861e17f472a1f44420408da6cefecb358b29bd4e8a5166746780c2"} Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.795902 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69897166-b5c9-404e-ac62-0b55f0960278","Type":"ContainerStarted","Data":"a1a8cae69793fcdeda0115c040ea95de4e540df38f817fe5d2600245f84cba4b"} Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.799757 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4559a88d-6d07-44d0-a829-c35fd773fcfb","Type":"ContainerStarted","Data":"9039d4893d7c154aacf5e3d7aa83396bed1a0000865ee2be9f2d19e6c72246c2"} Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.799926 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.802374 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hkgwr" event={"ID":"b1bdf176-776a-4211-8b43-ff90225de25a","Type":"ContainerStarted","Data":"4f561afbdd4fb1259076c496dea33b5413f902ebee541fb48eb7e1aa8fa45027"} Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.802487 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-qrkhk" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.802778 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-hkgwr" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.943451 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=17.021355905 podStartE2EDuration="26.943425708s" podCreationTimestamp="2025-11-28 11:41:46 +0000 UTC" firstStartedPulling="2025-11-28 11:42:01.505671274 +0000 UTC m=+1199.038185195" lastFinishedPulling="2025-11-28 11:42:11.427741067 +0000 UTC m=+1208.960254998" observedRunningTime="2025-11-28 11:42:12.936762551 +0000 UTC m=+1210.469276482" watchObservedRunningTime="2025-11-28 11:42:12.943425708 +0000 UTC m=+1210.475939629" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.947168 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-hkgwr" podStartSLOduration=15.88401714 podStartE2EDuration="23.947146207s" podCreationTimestamp="2025-11-28 11:41:49 +0000 UTC" firstStartedPulling="2025-11-28 11:42:01.628015102 +0000 UTC m=+1199.160529023" lastFinishedPulling="2025-11-28 11:42:09.691144159 +0000 UTC m=+1207.223658090" observedRunningTime="2025-11-28 11:42:12.921560828 +0000 UTC m=+1210.454074939" watchObservedRunningTime="2025-11-28 11:42:12.947146207 +0000 UTC m=+1210.479660128" Nov 28 11:42:12 crc kubenswrapper[4862]: I1128 11:42:12.987737 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=22.232712988 podStartE2EDuration="29.987717824s" podCreationTimestamp="2025-11-28 11:41:43 +0000 UTC" firstStartedPulling="2025-11-28 11:42:01.49835887 +0000 UTC m=+1199.030872791" lastFinishedPulling="2025-11-28 11:42:09.253363696 +0000 UTC m=+1206.785877627" observedRunningTime="2025-11-28 11:42:12.977422001 +0000 UTC m=+1210.509935922" watchObservedRunningTime="2025-11-28 11:42:12.987717824 +0000 UTC m=+1210.520231745" Nov 28 11:42:13 crc kubenswrapper[4862]: I1128 11:42:13.035200 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-qrkhk"] Nov 28 11:42:13 crc kubenswrapper[4862]: I1128 11:42:13.043798 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-qrkhk"] Nov 28 11:42:13 crc kubenswrapper[4862]: I1128 11:42:13.812618 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hdvv6" event={"ID":"ba7cc6b9-1443-45de-84f4-4ec3b92e699c","Type":"ContainerStarted","Data":"c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3"} Nov 28 11:42:13 crc kubenswrapper[4862]: I1128 11:42:13.812987 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hdvv6" event={"ID":"ba7cc6b9-1443-45de-84f4-4ec3b92e699c","Type":"ContainerStarted","Data":"49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8"} Nov 28 11:42:13 crc kubenswrapper[4862]: I1128 11:42:13.813784 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:42:13 crc kubenswrapper[4862]: I1128 11:42:13.840597 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-hdvv6" podStartSLOduration=16.845023985 podStartE2EDuration="24.840579438s" podCreationTimestamp="2025-11-28 11:41:49 +0000 UTC" firstStartedPulling="2025-11-28 11:42:01.597484381 +0000 UTC m=+1199.129998312" lastFinishedPulling="2025-11-28 11:42:09.593039844 +0000 UTC m=+1207.125553765" observedRunningTime="2025-11-28 11:42:13.833906831 +0000 UTC m=+1211.366420782" watchObservedRunningTime="2025-11-28 11:42:13.840579438 +0000 UTC m=+1211.373093349" Nov 28 11:42:14 crc kubenswrapper[4862]: I1128 11:42:14.825848 4862 generic.go:334] "Generic (PLEG): container finished" podID="27c60127-ee39-4d28-89f7-e5225ddc63d7" containerID="fcd0cc7dfd92fb5cac96b835cfe5b465117571bc62e3995b22db01397ac53877" exitCode=0 Nov 28 11:42:14 crc kubenswrapper[4862]: I1128 11:42:14.826887 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"27c60127-ee39-4d28-89f7-e5225ddc63d7","Type":"ContainerDied","Data":"fcd0cc7dfd92fb5cac96b835cfe5b465117571bc62e3995b22db01397ac53877"} Nov 28 11:42:14 crc kubenswrapper[4862]: I1128 11:42:14.827329 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:42:14 crc kubenswrapper[4862]: I1128 11:42:14.867605 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c23915a-f7a4-4334-aa70-31bf066f10d3" path="/var/lib/kubelet/pods/2c23915a-f7a4-4334-aa70-31bf066f10d3/volumes" Nov 28 11:42:15 crc kubenswrapper[4862]: I1128 11:42:15.840487 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"27c60127-ee39-4d28-89f7-e5225ddc63d7","Type":"ContainerStarted","Data":"36a419bb69dcce653bb57cc1d9e016f4dc4b3dc163be68616273503f06ee3dc5"} Nov 28 11:42:15 crc kubenswrapper[4862]: I1128 11:42:15.844032 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69897166-b5c9-404e-ac62-0b55f0960278","Type":"ContainerStarted","Data":"670f94a7e5c8ef75b6abff70cce614bc7b826f6dcb5baf65b275f5432887d508"} Nov 28 11:42:15 crc kubenswrapper[4862]: I1128 11:42:15.846440 4862 generic.go:334] "Generic (PLEG): container finished" podID="f01191bd-12ff-43ea-9788-381a997a3c08" containerID="7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755" exitCode=0 Nov 28 11:42:15 crc kubenswrapper[4862]: I1128 11:42:15.846541 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f01191bd-12ff-43ea-9788-381a997a3c08","Type":"ContainerDied","Data":"7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755"} Nov 28 11:42:15 crc kubenswrapper[4862]: I1128 11:42:15.849899 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b","Type":"ContainerStarted","Data":"e91e1d7ad1e2a8876f0520009135a4b0b1915dd5112109b61fca65064cefcd34"} Nov 28 11:42:15 crc kubenswrapper[4862]: I1128 11:42:15.863452 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=26.550665899 podStartE2EDuration="33.863428224s" podCreationTimestamp="2025-11-28 11:41:42 +0000 UTC" firstStartedPulling="2025-11-28 11:42:01.622350392 +0000 UTC m=+1199.154864333" lastFinishedPulling="2025-11-28 11:42:08.935112727 +0000 UTC m=+1206.467626658" observedRunningTime="2025-11-28 11:42:15.860833185 +0000 UTC m=+1213.393347136" watchObservedRunningTime="2025-11-28 11:42:15.863428224 +0000 UTC m=+1213.395942155" Nov 28 11:42:15 crc kubenswrapper[4862]: I1128 11:42:15.897475 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.178419966 podStartE2EDuration="24.897411677s" podCreationTimestamp="2025-11-28 11:41:51 +0000 UTC" firstStartedPulling="2025-11-28 11:42:01.717834307 +0000 UTC m=+1199.250348228" lastFinishedPulling="2025-11-28 11:42:15.436826018 +0000 UTC m=+1212.969339939" observedRunningTime="2025-11-28 11:42:15.883134978 +0000 UTC m=+1213.415648929" watchObservedRunningTime="2025-11-28 11:42:15.897411677 +0000 UTC m=+1213.429925638" Nov 28 11:42:15 crc kubenswrapper[4862]: I1128 11:42:15.932046 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=10.372824217 podStartE2EDuration="23.932021286s" podCreationTimestamp="2025-11-28 11:41:52 +0000 UTC" firstStartedPulling="2025-11-28 11:42:01.863984097 +0000 UTC m=+1199.396498018" lastFinishedPulling="2025-11-28 11:42:15.423181166 +0000 UTC m=+1212.955695087" observedRunningTime="2025-11-28 11:42:15.922243646 +0000 UTC m=+1213.454757607" watchObservedRunningTime="2025-11-28 11:42:15.932021286 +0000 UTC m=+1213.464535217" Nov 28 11:42:16 crc kubenswrapper[4862]: I1128 11:42:16.769473 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 28 11:42:16 crc kubenswrapper[4862]: I1128 11:42:16.811491 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 28 11:42:16 crc kubenswrapper[4862]: I1128 11:42:16.860186 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f01191bd-12ff-43ea-9788-381a997a3c08","Type":"ContainerStarted","Data":"94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980"} Nov 28 11:42:16 crc kubenswrapper[4862]: I1128 11:42:16.860969 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 28 11:42:16 crc kubenswrapper[4862]: I1128 11:42:16.902550 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=27.128290455 podStartE2EDuration="35.902524132s" podCreationTimestamp="2025-11-28 11:41:41 +0000 UTC" firstStartedPulling="2025-11-28 11:42:00.816350642 +0000 UTC m=+1198.348864563" lastFinishedPulling="2025-11-28 11:42:09.590584279 +0000 UTC m=+1207.123098240" observedRunningTime="2025-11-28 11:42:16.88929236 +0000 UTC m=+1214.421806301" watchObservedRunningTime="2025-11-28 11:42:16.902524132 +0000 UTC m=+1214.435038083" Nov 28 11:42:17 crc kubenswrapper[4862]: I1128 11:42:17.515474 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 28 11:42:17 crc kubenswrapper[4862]: I1128 11:42:17.597848 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 28 11:42:17 crc kubenswrapper[4862]: I1128 11:42:17.826649 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 28 11:42:17 crc kubenswrapper[4862]: I1128 11:42:17.890183 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"527b2bce-1186-4186-8992-a3a63d2d2b22","Type":"ContainerStarted","Data":"b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2"} Nov 28 11:42:17 crc kubenswrapper[4862]: I1128 11:42:17.891368 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 28 11:42:17 crc kubenswrapper[4862]: I1128 11:42:17.969730 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.102854 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fbf586c4f-2n9jk"] Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.104029 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.105949 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.117672 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fbf586c4f-2n9jk"] Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.161033 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-flbld"] Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.162036 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.164746 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.185572 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-flbld"] Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224189 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovs-rundir\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224234 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-dns-svc\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224272 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-ovsdbserver-sb\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224303 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-config\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224342 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224384 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovn-rundir\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224403 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56qcf\" (UniqueName: \"kubernetes.io/projected/53d62b74-0406-47fe-8a9a-36cc6f288f27-kube-api-access-56qcf\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224421 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm8jf\" (UniqueName: \"kubernetes.io/projected/1de81805-45d6-4b79-870b-da707ac5b86c-kube-api-access-vm8jf\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224445 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53d62b74-0406-47fe-8a9a-36cc6f288f27-config\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.224468 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-combined-ca-bundle\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.244691 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fbf586c4f-2n9jk"] Nov 28 11:42:18 crc kubenswrapper[4862]: E1128 11:42:18.245261 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-vm8jf ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" podUID="1de81805-45d6-4b79-870b-da707ac5b86c" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.271838 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58bd875f97-668vf"] Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.279475 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.281130 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58bd875f97-668vf"] Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.283247 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.312993 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.314367 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.316763 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.316956 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-bcftp" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.317151 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.317312 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.325404 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgllg\" (UniqueName: \"kubernetes.io/projected/91d6e5f5-d164-4e46-8e3b-ca0357644965-kube-api-access-bgllg\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.325475 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovn-rundir\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.325497 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56qcf\" (UniqueName: \"kubernetes.io/projected/53d62b74-0406-47fe-8a9a-36cc6f288f27-kube-api-access-56qcf\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.325512 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm8jf\" (UniqueName: \"kubernetes.io/projected/1de81805-45d6-4b79-870b-da707ac5b86c-kube-api-access-vm8jf\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.325538 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-dns-svc\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.325559 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53d62b74-0406-47fe-8a9a-36cc6f288f27-config\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.325852 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovn-rundir\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326694 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53d62b74-0406-47fe-8a9a-36cc6f288f27-config\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326751 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-combined-ca-bundle\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326783 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovs-rundir\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326802 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-config\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326821 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-nb\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326841 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-dns-svc\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326876 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-ovsdbserver-sb\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326908 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-config\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326932 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-sb\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.326963 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.330279 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-ovsdbserver-sb\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.330373 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovs-rundir\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.330596 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-dns-svc\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.330951 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-config\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.334883 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-combined-ca-bundle\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.335344 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.338050 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.348972 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56qcf\" (UniqueName: \"kubernetes.io/projected/53d62b74-0406-47fe-8a9a-36cc6f288f27-kube-api-access-56qcf\") pod \"ovn-controller-metrics-flbld\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.350132 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm8jf\" (UniqueName: \"kubernetes.io/projected/1de81805-45d6-4b79-870b-da707ac5b86c-kube-api-access-vm8jf\") pod \"dnsmasq-dns-fbf586c4f-2n9jk\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.427936 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.427992 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-dns-svc\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428027 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stjnh\" (UniqueName: \"kubernetes.io/projected/5e37bdc4-788e-4889-9de7-682e77c3c5eb-kube-api-access-stjnh\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428048 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-config\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428069 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-nb\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428124 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-config\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428150 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428170 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-sb\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428187 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428214 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-scripts\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428233 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgllg\" (UniqueName: \"kubernetes.io/projected/91d6e5f5-d164-4e46-8e3b-ca0357644965-kube-api-access-bgllg\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.428258 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.429058 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-dns-svc\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.429575 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-config\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.430060 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-nb\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.430844 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-sb\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.445195 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgllg\" (UniqueName: \"kubernetes.io/projected/91d6e5f5-d164-4e46-8e3b-ca0357644965-kube-api-access-bgllg\") pod \"dnsmasq-dns-58bd875f97-668vf\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.483128 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.529764 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stjnh\" (UniqueName: \"kubernetes.io/projected/5e37bdc4-788e-4889-9de7-682e77c3c5eb-kube-api-access-stjnh\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.529894 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-config\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.529955 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.529992 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.530049 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-scripts\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.530119 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.530179 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.530614 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.531114 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-scripts\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.531272 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-config\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.539768 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.539843 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.539922 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.549773 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stjnh\" (UniqueName: \"kubernetes.io/projected/5e37bdc4-788e-4889-9de7-682e77c3c5eb-kube-api-access-stjnh\") pod \"ovn-northd-0\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.600789 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.710231 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.898158 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.917333 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.920685 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-flbld"] Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.938693 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-dns-svc\") pod \"1de81805-45d6-4b79-870b-da707ac5b86c\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.938895 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vm8jf\" (UniqueName: \"kubernetes.io/projected/1de81805-45d6-4b79-870b-da707ac5b86c-kube-api-access-vm8jf\") pod \"1de81805-45d6-4b79-870b-da707ac5b86c\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.938951 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-config\") pod \"1de81805-45d6-4b79-870b-da707ac5b86c\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.938982 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-ovsdbserver-sb\") pod \"1de81805-45d6-4b79-870b-da707ac5b86c\" (UID: \"1de81805-45d6-4b79-870b-da707ac5b86c\") " Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.939578 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1de81805-45d6-4b79-870b-da707ac5b86c" (UID: "1de81805-45d6-4b79-870b-da707ac5b86c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.939718 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1de81805-45d6-4b79-870b-da707ac5b86c" (UID: "1de81805-45d6-4b79-870b-da707ac5b86c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.939752 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-config" (OuterVolumeSpecName: "config") pod "1de81805-45d6-4b79-870b-da707ac5b86c" (UID: "1de81805-45d6-4b79-870b-da707ac5b86c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.940113 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.940603 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.940649 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de81805-45d6-4b79-870b-da707ac5b86c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:18 crc kubenswrapper[4862]: I1128 11:42:18.942930 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de81805-45d6-4b79-870b-da707ac5b86c-kube-api-access-vm8jf" (OuterVolumeSpecName: "kube-api-access-vm8jf") pod "1de81805-45d6-4b79-870b-da707ac5b86c" (UID: "1de81805-45d6-4b79-870b-da707ac5b86c"). InnerVolumeSpecName "kube-api-access-vm8jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.021145 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58bd875f97-668vf"] Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.044075 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vm8jf\" (UniqueName: \"kubernetes.io/projected/1de81805-45d6-4b79-870b-da707ac5b86c-kube-api-access-vm8jf\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:19 crc kubenswrapper[4862]: W1128 11:42:19.133305 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e37bdc4_788e_4889_9de7_682e77c3c5eb.slice/crio-8d2a830fa44e976a4feb94ac2049af8d2b090df8bf4e0247a72e920f00732e06 WatchSource:0}: Error finding container 8d2a830fa44e976a4feb94ac2049af8d2b090df8bf4e0247a72e920f00732e06: Status 404 returned error can't find the container with id 8d2a830fa44e976a4feb94ac2049af8d2b090df8bf4e0247a72e920f00732e06 Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.135379 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.227078 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.909482 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5e37bdc4-788e-4889-9de7-682e77c3c5eb","Type":"ContainerStarted","Data":"8d2a830fa44e976a4feb94ac2049af8d2b090df8bf4e0247a72e920f00732e06"} Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.911223 4862 generic.go:334] "Generic (PLEG): container finished" podID="91d6e5f5-d164-4e46-8e3b-ca0357644965" containerID="cc452e3fb71d1ec3392cb8f7cbffebee7bf2eb82ad062a5b6281c72bb35909b3" exitCode=0 Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.911351 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd875f97-668vf" event={"ID":"91d6e5f5-d164-4e46-8e3b-ca0357644965","Type":"ContainerDied","Data":"cc452e3fb71d1ec3392cb8f7cbffebee7bf2eb82ad062a5b6281c72bb35909b3"} Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.911475 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd875f97-668vf" event={"ID":"91d6e5f5-d164-4e46-8e3b-ca0357644965","Type":"ContainerStarted","Data":"4b8c634c432992d209d1c38cb174b6c73aa16341422e80da6d953dfd61a2da68"} Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.912769 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-flbld" event={"ID":"53d62b74-0406-47fe-8a9a-36cc6f288f27","Type":"ContainerStarted","Data":"a1e1c2cca57dec29a126aac011c043be5314d2cdd1005a224a53029e52b33e16"} Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.912798 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-flbld" event={"ID":"53d62b74-0406-47fe-8a9a-36cc6f288f27","Type":"ContainerStarted","Data":"29b56b3656a530dddc3c588840c5692f6ddc8b025edb666c85995871c468e387"} Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.912870 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbf586c4f-2n9jk" Nov 28 11:42:19 crc kubenswrapper[4862]: I1128 11:42:19.982782 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-flbld" podStartSLOduration=1.9827622329999999 podStartE2EDuration="1.982762233s" podCreationTimestamp="2025-11-28 11:42:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:42:19.979291741 +0000 UTC m=+1217.511805662" watchObservedRunningTime="2025-11-28 11:42:19.982762233 +0000 UTC m=+1217.515276164" Nov 28 11:42:20 crc kubenswrapper[4862]: I1128 11:42:20.071423 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fbf586c4f-2n9jk"] Nov 28 11:42:20 crc kubenswrapper[4862]: I1128 11:42:20.078664 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fbf586c4f-2n9jk"] Nov 28 11:42:20 crc kubenswrapper[4862]: I1128 11:42:20.859635 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de81805-45d6-4b79-870b-da707ac5b86c" path="/var/lib/kubelet/pods/1de81805-45d6-4b79-870b-da707ac5b86c/volumes" Nov 28 11:42:20 crc kubenswrapper[4862]: I1128 11:42:20.925277 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd875f97-668vf" event={"ID":"91d6e5f5-d164-4e46-8e3b-ca0357644965","Type":"ContainerStarted","Data":"c73274ea7625201b10973ff30d6da6910e4eafb995e62264aca75fc38d9b6e74"} Nov 28 11:42:20 crc kubenswrapper[4862]: I1128 11:42:20.961530 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58bd875f97-668vf" podStartSLOduration=2.961510649 podStartE2EDuration="2.961510649s" podCreationTimestamp="2025-11-28 11:42:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:42:20.947280281 +0000 UTC m=+1218.479794222" watchObservedRunningTime="2025-11-28 11:42:20.961510649 +0000 UTC m=+1218.494024570" Nov 28 11:42:21 crc kubenswrapper[4862]: I1128 11:42:21.937219 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5e37bdc4-788e-4889-9de7-682e77c3c5eb","Type":"ContainerStarted","Data":"23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3"} Nov 28 11:42:21 crc kubenswrapper[4862]: I1128 11:42:21.937953 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5e37bdc4-788e-4889-9de7-682e77c3c5eb","Type":"ContainerStarted","Data":"3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6"} Nov 28 11:42:21 crc kubenswrapper[4862]: I1128 11:42:21.937988 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:21 crc kubenswrapper[4862]: I1128 11:42:21.938012 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 28 11:42:21 crc kubenswrapper[4862]: I1128 11:42:21.965972 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.044063071 podStartE2EDuration="3.965950057s" podCreationTimestamp="2025-11-28 11:42:18 +0000 UTC" firstStartedPulling="2025-11-28 11:42:19.137032439 +0000 UTC m=+1216.669546360" lastFinishedPulling="2025-11-28 11:42:21.058919415 +0000 UTC m=+1218.591433346" observedRunningTime="2025-11-28 11:42:21.958019077 +0000 UTC m=+1219.490532998" watchObservedRunningTime="2025-11-28 11:42:21.965950057 +0000 UTC m=+1219.498463988" Nov 28 11:42:22 crc kubenswrapper[4862]: I1128 11:42:22.478785 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 28 11:42:22 crc kubenswrapper[4862]: I1128 11:42:22.478851 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 28 11:42:22 crc kubenswrapper[4862]: I1128 11:42:22.615671 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.063389 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.853239 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bc3a-account-create-update-x9czv"] Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.855247 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.857791 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.872060 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bc3a-account-create-update-x9czv"] Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.914747 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-62jdk"] Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.915858 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.923552 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-62jdk"] Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.981055 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-operator-scripts\") pod \"keystone-db-create-62jdk\" (UID: \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\") " pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.981266 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/705faed3-031b-4ec7-bfd7-b5ac626c98b3-operator-scripts\") pod \"keystone-bc3a-account-create-update-x9czv\" (UID: \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\") " pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.981299 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fhtb\" (UniqueName: \"kubernetes.io/projected/705faed3-031b-4ec7-bfd7-b5ac626c98b3-kube-api-access-4fhtb\") pod \"keystone-bc3a-account-create-update-x9czv\" (UID: \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\") " pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:23 crc kubenswrapper[4862]: I1128 11:42:23.981503 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plzzf\" (UniqueName: \"kubernetes.io/projected/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-kube-api-access-plzzf\") pod \"keystone-db-create-62jdk\" (UID: \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\") " pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.015539 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.015822 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.083480 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-operator-scripts\") pod \"keystone-db-create-62jdk\" (UID: \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\") " pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.083666 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/705faed3-031b-4ec7-bfd7-b5ac626c98b3-operator-scripts\") pod \"keystone-bc3a-account-create-update-x9czv\" (UID: \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\") " pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.083697 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fhtb\" (UniqueName: \"kubernetes.io/projected/705faed3-031b-4ec7-bfd7-b5ac626c98b3-kube-api-access-4fhtb\") pod \"keystone-bc3a-account-create-update-x9czv\" (UID: \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\") " pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.083777 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plzzf\" (UniqueName: \"kubernetes.io/projected/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-kube-api-access-plzzf\") pod \"keystone-db-create-62jdk\" (UID: \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\") " pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.084520 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-operator-scripts\") pod \"keystone-db-create-62jdk\" (UID: \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\") " pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.085283 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/705faed3-031b-4ec7-bfd7-b5ac626c98b3-operator-scripts\") pod \"keystone-bc3a-account-create-update-x9czv\" (UID: \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\") " pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.114763 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fhtb\" (UniqueName: \"kubernetes.io/projected/705faed3-031b-4ec7-bfd7-b5ac626c98b3-kube-api-access-4fhtb\") pod \"keystone-bc3a-account-create-update-x9czv\" (UID: \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\") " pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.114925 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.115328 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plzzf\" (UniqueName: \"kubernetes.io/projected/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-kube-api-access-plzzf\") pod \"keystone-db-create-62jdk\" (UID: \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\") " pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.185034 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.187806 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-q4jzk"] Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.189036 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.221957 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q4jzk"] Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.230799 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-e4ca-account-create-update-hsh77"] Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.232486 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.231282 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.239832 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.243814 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e4ca-account-create-update-hsh77"] Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.289179 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxwvr\" (UniqueName: \"kubernetes.io/projected/d990f484-bffa-4bfc-8c45-b00c5514512a-kube-api-access-sxwvr\") pod \"placement-e4ca-account-create-update-hsh77\" (UID: \"d990f484-bffa-4bfc-8c45-b00c5514512a\") " pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.289245 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec7a6dc-a28c-4239-b112-6139ae089cee-operator-scripts\") pod \"placement-db-create-q4jzk\" (UID: \"1ec7a6dc-a28c-4239-b112-6139ae089cee\") " pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.289312 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d990f484-bffa-4bfc-8c45-b00c5514512a-operator-scripts\") pod \"placement-e4ca-account-create-update-hsh77\" (UID: \"d990f484-bffa-4bfc-8c45-b00c5514512a\") " pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.289364 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b89v6\" (UniqueName: \"kubernetes.io/projected/1ec7a6dc-a28c-4239-b112-6139ae089cee-kube-api-access-b89v6\") pod \"placement-db-create-q4jzk\" (UID: \"1ec7a6dc-a28c-4239-b112-6139ae089cee\") " pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.391268 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxwvr\" (UniqueName: \"kubernetes.io/projected/d990f484-bffa-4bfc-8c45-b00c5514512a-kube-api-access-sxwvr\") pod \"placement-e4ca-account-create-update-hsh77\" (UID: \"d990f484-bffa-4bfc-8c45-b00c5514512a\") " pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.391341 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec7a6dc-a28c-4239-b112-6139ae089cee-operator-scripts\") pod \"placement-db-create-q4jzk\" (UID: \"1ec7a6dc-a28c-4239-b112-6139ae089cee\") " pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.391419 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d990f484-bffa-4bfc-8c45-b00c5514512a-operator-scripts\") pod \"placement-e4ca-account-create-update-hsh77\" (UID: \"d990f484-bffa-4bfc-8c45-b00c5514512a\") " pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.391444 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b89v6\" (UniqueName: \"kubernetes.io/projected/1ec7a6dc-a28c-4239-b112-6139ae089cee-kube-api-access-b89v6\") pod \"placement-db-create-q4jzk\" (UID: \"1ec7a6dc-a28c-4239-b112-6139ae089cee\") " pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.392503 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec7a6dc-a28c-4239-b112-6139ae089cee-operator-scripts\") pod \"placement-db-create-q4jzk\" (UID: \"1ec7a6dc-a28c-4239-b112-6139ae089cee\") " pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.392604 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d990f484-bffa-4bfc-8c45-b00c5514512a-operator-scripts\") pod \"placement-e4ca-account-create-update-hsh77\" (UID: \"d990f484-bffa-4bfc-8c45-b00c5514512a\") " pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.426409 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b89v6\" (UniqueName: \"kubernetes.io/projected/1ec7a6dc-a28c-4239-b112-6139ae089cee-kube-api-access-b89v6\") pod \"placement-db-create-q4jzk\" (UID: \"1ec7a6dc-a28c-4239-b112-6139ae089cee\") " pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.427514 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxwvr\" (UniqueName: \"kubernetes.io/projected/d990f484-bffa-4bfc-8c45-b00c5514512a-kube-api-access-sxwvr\") pod \"placement-e4ca-account-create-update-hsh77\" (UID: \"d990f484-bffa-4bfc-8c45-b00c5514512a\") " pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.628492 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.634893 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.868230 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-62jdk"] Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.880990 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bc3a-account-create-update-x9czv"] Nov 28 11:42:24 crc kubenswrapper[4862]: W1128 11:42:24.889580 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod705faed3_031b_4ec7_bfd7_b5ac626c98b3.slice/crio-71891aebfed6dbf2274befd359365a530000a88a41286ec99c76ef3fbb1ec650 WatchSource:0}: Error finding container 71891aebfed6dbf2274befd359365a530000a88a41286ec99c76ef3fbb1ec650: Status 404 returned error can't find the container with id 71891aebfed6dbf2274befd359365a530000a88a41286ec99c76ef3fbb1ec650 Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.979874 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bc3a-account-create-update-x9czv" event={"ID":"705faed3-031b-4ec7-bfd7-b5ac626c98b3","Type":"ContainerStarted","Data":"71891aebfed6dbf2274befd359365a530000a88a41286ec99c76ef3fbb1ec650"} Nov 28 11:42:24 crc kubenswrapper[4862]: I1128 11:42:24.981976 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-62jdk" event={"ID":"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527","Type":"ContainerStarted","Data":"d4ee9f46c77bd2298c01533923bc21b6f15de63f6050bacea1bd3ae564e59a41"} Nov 28 11:42:25 crc kubenswrapper[4862]: I1128 11:42:25.067950 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 28 11:42:25 crc kubenswrapper[4862]: I1128 11:42:25.189795 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q4jzk"] Nov 28 11:42:25 crc kubenswrapper[4862]: I1128 11:42:25.233289 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e4ca-account-create-update-hsh77"] Nov 28 11:42:25 crc kubenswrapper[4862]: W1128 11:42:25.237233 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd990f484_bffa_4bfc_8c45_b00c5514512a.slice/crio-6ced57fa1cbe12c8075eb1964a75a704c9c38e6f6af5b27dceab9c29c11c88f5 WatchSource:0}: Error finding container 6ced57fa1cbe12c8075eb1964a75a704c9c38e6f6af5b27dceab9c29c11c88f5: Status 404 returned error can't find the container with id 6ced57fa1cbe12c8075eb1964a75a704c9c38e6f6af5b27dceab9c29c11c88f5 Nov 28 11:42:25 crc kubenswrapper[4862]: I1128 11:42:25.992315 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q4jzk" event={"ID":"1ec7a6dc-a28c-4239-b112-6139ae089cee","Type":"ContainerStarted","Data":"784fefb4c2627b19d12af7de42886cb1b9f414241d8468f68a9bc666c0d14a1f"} Nov 28 11:42:25 crc kubenswrapper[4862]: I1128 11:42:25.993907 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e4ca-account-create-update-hsh77" event={"ID":"d990f484-bffa-4bfc-8c45-b00c5514512a","Type":"ContainerStarted","Data":"6ced57fa1cbe12c8075eb1964a75a704c9c38e6f6af5b27dceab9c29c11c88f5"} Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.474427 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.583872 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58bd875f97-668vf"] Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.584112 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58bd875f97-668vf" podUID="91d6e5f5-d164-4e46-8e3b-ca0357644965" containerName="dnsmasq-dns" containerID="cri-o://c73274ea7625201b10973ff30d6da6910e4eafb995e62264aca75fc38d9b6e74" gracePeriod=10 Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.592107 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.627801 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c8cb8df65-lvvcw"] Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.629122 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.659877 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c8cb8df65-lvvcw"] Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.755168 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.755221 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.755244 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-config\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.755476 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-dns-svc\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.755518 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd92k\" (UniqueName: \"kubernetes.io/projected/966949c3-c79d-4811-8c75-c4e913cd6a19-kube-api-access-bd92k\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.857444 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.857498 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-config\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.857580 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-dns-svc\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.857605 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd92k\" (UniqueName: \"kubernetes.io/projected/966949c3-c79d-4811-8c75-c4e913cd6a19-kube-api-access-bd92k\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.858361 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.858412 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-config\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.858498 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-dns-svc\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.858827 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.859330 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.886061 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd92k\" (UniqueName: \"kubernetes.io/projected/966949c3-c79d-4811-8c75-c4e913cd6a19-kube-api-access-bd92k\") pod \"dnsmasq-dns-7c8cb8df65-lvvcw\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:26 crc kubenswrapper[4862]: I1128 11:42:26.967312 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.005341 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bc3a-account-create-update-x9czv" event={"ID":"705faed3-031b-4ec7-bfd7-b5ac626c98b3","Type":"ContainerStarted","Data":"e461f0dccb2ed908a6c4512b4a8acd9a84d85b485f10f9996cba1d1617b20a5d"} Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.008577 4862 generic.go:334] "Generic (PLEG): container finished" podID="a51e2e8d-a77a-4017-90bd-1e2dfdc8f527" containerID="aedd0f47afe828a0180407f4e6ad432542ca5f49e5f4367abad5584e65aa27c7" exitCode=0 Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.008844 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-62jdk" event={"ID":"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527","Type":"ContainerDied","Data":"aedd0f47afe828a0180407f4e6ad432542ca5f49e5f4367abad5584e65aa27c7"} Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.011630 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e4ca-account-create-update-hsh77" event={"ID":"d990f484-bffa-4bfc-8c45-b00c5514512a","Type":"ContainerStarted","Data":"9c34d5f4b793116c2cc68fd196f337aca1ef159640d7e85b2d3fbe1065f54a98"} Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.013402 4862 generic.go:334] "Generic (PLEG): container finished" podID="91d6e5f5-d164-4e46-8e3b-ca0357644965" containerID="c73274ea7625201b10973ff30d6da6910e4eafb995e62264aca75fc38d9b6e74" exitCode=0 Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.013461 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd875f97-668vf" event={"ID":"91d6e5f5-d164-4e46-8e3b-ca0357644965","Type":"ContainerDied","Data":"c73274ea7625201b10973ff30d6da6910e4eafb995e62264aca75fc38d9b6e74"} Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.016011 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q4jzk" event={"ID":"1ec7a6dc-a28c-4239-b112-6139ae089cee","Type":"ContainerStarted","Data":"1b4d99bd2494049894725880aae78e315feeedf9ec2f86b9e684dcb2edada6c1"} Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.025927 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bc3a-account-create-update-x9czv" podStartSLOduration=4.025906169 podStartE2EDuration="4.025906169s" podCreationTimestamp="2025-11-28 11:42:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:42:27.024024209 +0000 UTC m=+1224.556538130" watchObservedRunningTime="2025-11-28 11:42:27.025906169 +0000 UTC m=+1224.558420090" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.047332 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-e4ca-account-create-update-hsh77" podStartSLOduration=3.047312517 podStartE2EDuration="3.047312517s" podCreationTimestamp="2025-11-28 11:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:42:27.043119476 +0000 UTC m=+1224.575633397" watchObservedRunningTime="2025-11-28 11:42:27.047312517 +0000 UTC m=+1224.579826438" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.058775 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-q4jzk" podStartSLOduration=3.058758101 podStartE2EDuration="3.058758101s" podCreationTimestamp="2025-11-28 11:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:42:27.055739561 +0000 UTC m=+1224.588253482" watchObservedRunningTime="2025-11-28 11:42:27.058758101 +0000 UTC m=+1224.591272022" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.443680 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c8cb8df65-lvvcw"] Nov 28 11:42:27 crc kubenswrapper[4862]: W1128 11:42:27.456005 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod966949c3_c79d_4811_8c75_c4e913cd6a19.slice/crio-2b6ea71ea113385a11cc58c86cddd888e3de762caa0ed7a8592c00822a2db0a3 WatchSource:0}: Error finding container 2b6ea71ea113385a11cc58c86cddd888e3de762caa0ed7a8592c00822a2db0a3: Status 404 returned error can't find the container with id 2b6ea71ea113385a11cc58c86cddd888e3de762caa0ed7a8592c00822a2db0a3 Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.525356 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.573370 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-sb\") pod \"91d6e5f5-d164-4e46-8e3b-ca0357644965\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.573760 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-config\") pod \"91d6e5f5-d164-4e46-8e3b-ca0357644965\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.574024 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-nb\") pod \"91d6e5f5-d164-4e46-8e3b-ca0357644965\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.574174 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgllg\" (UniqueName: \"kubernetes.io/projected/91d6e5f5-d164-4e46-8e3b-ca0357644965-kube-api-access-bgllg\") pod \"91d6e5f5-d164-4e46-8e3b-ca0357644965\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.574299 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-dns-svc\") pod \"91d6e5f5-d164-4e46-8e3b-ca0357644965\" (UID: \"91d6e5f5-d164-4e46-8e3b-ca0357644965\") " Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.578787 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91d6e5f5-d164-4e46-8e3b-ca0357644965-kube-api-access-bgllg" (OuterVolumeSpecName: "kube-api-access-bgllg") pod "91d6e5f5-d164-4e46-8e3b-ca0357644965" (UID: "91d6e5f5-d164-4e46-8e3b-ca0357644965"). InnerVolumeSpecName "kube-api-access-bgllg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.668869 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "91d6e5f5-d164-4e46-8e3b-ca0357644965" (UID: "91d6e5f5-d164-4e46-8e3b-ca0357644965"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.670389 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "91d6e5f5-d164-4e46-8e3b-ca0357644965" (UID: "91d6e5f5-d164-4e46-8e3b-ca0357644965"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.677414 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgllg\" (UniqueName: \"kubernetes.io/projected/91d6e5f5-d164-4e46-8e3b-ca0357644965-kube-api-access-bgllg\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.677459 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.677477 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.701657 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "91d6e5f5-d164-4e46-8e3b-ca0357644965" (UID: "91d6e5f5-d164-4e46-8e3b-ca0357644965"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.703681 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-config" (OuterVolumeSpecName: "config") pod "91d6e5f5-d164-4e46-8e3b-ca0357644965" (UID: "91d6e5f5-d164-4e46-8e3b-ca0357644965"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.709519 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 28 11:42:27 crc kubenswrapper[4862]: E1128 11:42:27.712818 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d6e5f5-d164-4e46-8e3b-ca0357644965" containerName="dnsmasq-dns" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.712843 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d6e5f5-d164-4e46-8e3b-ca0357644965" containerName="dnsmasq-dns" Nov 28 11:42:27 crc kubenswrapper[4862]: E1128 11:42:27.712868 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d6e5f5-d164-4e46-8e3b-ca0357644965" containerName="init" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.712877 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d6e5f5-d164-4e46-8e3b-ca0357644965" containerName="init" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.713058 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d6e5f5-d164-4e46-8e3b-ca0357644965" containerName="dnsmasq-dns" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.721928 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.722174 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.725025 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.725352 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.725046 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.725261 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-ftcjd" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.778749 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9lp4\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-kube-api-access-l9lp4\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.779151 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.779246 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-lock\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.779443 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-cache\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.779562 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.779636 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.779652 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91d6e5f5-d164-4e46-8e3b-ca0357644965-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.885161 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9lp4\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-kube-api-access-l9lp4\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.885431 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.885512 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-lock\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: E1128 11:42:27.885616 4862 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 11:42:27 crc kubenswrapper[4862]: E1128 11:42:27.885644 4862 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 11:42:27 crc kubenswrapper[4862]: E1128 11:42:27.885687 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift podName:b56372b7-fc39-40cc-8e0a-e95f907db697 nodeName:}" failed. No retries permitted until 2025-11-28 11:42:28.385668426 +0000 UTC m=+1225.918182347 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift") pod "swift-storage-0" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697") : configmap "swift-ring-files" not found Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.885803 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-cache\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.885931 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.886258 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.886430 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-cache\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.886428 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-lock\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.916064 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9lp4\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-kube-api-access-l9lp4\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:27 crc kubenswrapper[4862]: I1128 11:42:27.918720 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.026194 4862 generic.go:334] "Generic (PLEG): container finished" podID="966949c3-c79d-4811-8c75-c4e913cd6a19" containerID="ce8880d10b828a26bff152b7cdfba1672915a6ed5d72b43cd17a128c1264bfc5" exitCode=0 Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.026250 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" event={"ID":"966949c3-c79d-4811-8c75-c4e913cd6a19","Type":"ContainerDied","Data":"ce8880d10b828a26bff152b7cdfba1672915a6ed5d72b43cd17a128c1264bfc5"} Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.026275 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" event={"ID":"966949c3-c79d-4811-8c75-c4e913cd6a19","Type":"ContainerStarted","Data":"2b6ea71ea113385a11cc58c86cddd888e3de762caa0ed7a8592c00822a2db0a3"} Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.029882 4862 generic.go:334] "Generic (PLEG): container finished" podID="d990f484-bffa-4bfc-8c45-b00c5514512a" containerID="9c34d5f4b793116c2cc68fd196f337aca1ef159640d7e85b2d3fbe1065f54a98" exitCode=0 Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.030053 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e4ca-account-create-update-hsh77" event={"ID":"d990f484-bffa-4bfc-8c45-b00c5514512a","Type":"ContainerDied","Data":"9c34d5f4b793116c2cc68fd196f337aca1ef159640d7e85b2d3fbe1065f54a98"} Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.033269 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd875f97-668vf" event={"ID":"91d6e5f5-d164-4e46-8e3b-ca0357644965","Type":"ContainerDied","Data":"4b8c634c432992d209d1c38cb174b6c73aa16341422e80da6d953dfd61a2da68"} Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.033288 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bd875f97-668vf" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.033487 4862 scope.go:117] "RemoveContainer" containerID="c73274ea7625201b10973ff30d6da6910e4eafb995e62264aca75fc38d9b6e74" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.038857 4862 generic.go:334] "Generic (PLEG): container finished" podID="1ec7a6dc-a28c-4239-b112-6139ae089cee" containerID="1b4d99bd2494049894725880aae78e315feeedf9ec2f86b9e684dcb2edada6c1" exitCode=0 Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.038937 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q4jzk" event={"ID":"1ec7a6dc-a28c-4239-b112-6139ae089cee","Type":"ContainerDied","Data":"1b4d99bd2494049894725880aae78e315feeedf9ec2f86b9e684dcb2edada6c1"} Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.049957 4862 generic.go:334] "Generic (PLEG): container finished" podID="705faed3-031b-4ec7-bfd7-b5ac626c98b3" containerID="e461f0dccb2ed908a6c4512b4a8acd9a84d85b485f10f9996cba1d1617b20a5d" exitCode=0 Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.050375 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bc3a-account-create-update-x9czv" event={"ID":"705faed3-031b-4ec7-bfd7-b5ac626c98b3","Type":"ContainerDied","Data":"e461f0dccb2ed908a6c4512b4a8acd9a84d85b485f10f9996cba1d1617b20a5d"} Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.131753 4862 scope.go:117] "RemoveContainer" containerID="cc452e3fb71d1ec3392cb8f7cbffebee7bf2eb82ad062a5b6281c72bb35909b3" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.152599 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58bd875f97-668vf"] Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.188547 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58bd875f97-668vf"] Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.220215 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-xchsv"] Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.221380 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.223322 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.223452 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.223533 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.239985 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xchsv"] Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.303493 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-swiftconf\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.303686 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-ring-data-devices\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.303961 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-combined-ca-bundle\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.304344 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kklvf\" (UniqueName: \"kubernetes.io/projected/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-kube-api-access-kklvf\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.304411 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-scripts\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.304441 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-dispersionconf\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.304493 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-etc-swift\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.305311 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.405659 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-operator-scripts\") pod \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\" (UID: \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\") " Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.405801 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plzzf\" (UniqueName: \"kubernetes.io/projected/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-kube-api-access-plzzf\") pod \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\" (UID: \"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527\") " Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.406151 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.406185 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kklvf\" (UniqueName: \"kubernetes.io/projected/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-kube-api-access-kklvf\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.406211 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-scripts\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.406229 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-dispersionconf\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.406253 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-etc-swift\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.406330 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-ring-data-devices\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.406351 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-swiftconf\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.406399 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-combined-ca-bundle\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.406837 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a51e2e8d-a77a-4017-90bd-1e2dfdc8f527" (UID: "a51e2e8d-a77a-4017-90bd-1e2dfdc8f527"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:28 crc kubenswrapper[4862]: E1128 11:42:28.407126 4862 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 11:42:28 crc kubenswrapper[4862]: E1128 11:42:28.407157 4862 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 11:42:28 crc kubenswrapper[4862]: E1128 11:42:28.407211 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift podName:b56372b7-fc39-40cc-8e0a-e95f907db697 nodeName:}" failed. No retries permitted until 2025-11-28 11:42:29.407191573 +0000 UTC m=+1226.939705564 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift") pod "swift-storage-0" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697") : configmap "swift-ring-files" not found Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.407488 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-etc-swift\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.407659 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-scripts\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.408765 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-ring-data-devices\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.411072 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-dispersionconf\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.411071 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-kube-api-access-plzzf" (OuterVolumeSpecName: "kube-api-access-plzzf") pod "a51e2e8d-a77a-4017-90bd-1e2dfdc8f527" (UID: "a51e2e8d-a77a-4017-90bd-1e2dfdc8f527"). InnerVolumeSpecName "kube-api-access-plzzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.411865 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-combined-ca-bundle\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.412722 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-swiftconf\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.423311 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kklvf\" (UniqueName: \"kubernetes.io/projected/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-kube-api-access-kklvf\") pod \"swift-ring-rebalance-xchsv\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.508521 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.508586 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plzzf\" (UniqueName: \"kubernetes.io/projected/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527-kube-api-access-plzzf\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.544589 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.801863 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xchsv"] Nov 28 11:42:28 crc kubenswrapper[4862]: W1128 11:42:28.809991 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ac70dca_3c3a_43fe_81ef_2677c476e9a4.slice/crio-21261a8d996fd1db9b549697ba9c74193b4ad2804231cbb8e9f0cee7e36fb3ad WatchSource:0}: Error finding container 21261a8d996fd1db9b549697ba9c74193b4ad2804231cbb8e9f0cee7e36fb3ad: Status 404 returned error can't find the container with id 21261a8d996fd1db9b549697ba9c74193b4ad2804231cbb8e9f0cee7e36fb3ad Nov 28 11:42:28 crc kubenswrapper[4862]: I1128 11:42:28.869573 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91d6e5f5-d164-4e46-8e3b-ca0357644965" path="/var/lib/kubelet/pods/91d6e5f5-d164-4e46-8e3b-ca0357644965/volumes" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.069811 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xchsv" event={"ID":"3ac70dca-3c3a-43fe-81ef-2677c476e9a4","Type":"ContainerStarted","Data":"21261a8d996fd1db9b549697ba9c74193b4ad2804231cbb8e9f0cee7e36fb3ad"} Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.078512 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-62jdk" event={"ID":"a51e2e8d-a77a-4017-90bd-1e2dfdc8f527","Type":"ContainerDied","Data":"d4ee9f46c77bd2298c01533923bc21b6f15de63f6050bacea1bd3ae564e59a41"} Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.078584 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4ee9f46c77bd2298c01533923bc21b6f15de63f6050bacea1bd3ae564e59a41" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.078724 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-62jdk" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.116373 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" event={"ID":"966949c3-c79d-4811-8c75-c4e913cd6a19","Type":"ContainerStarted","Data":"97c0b76d1ae2e41b6a809bf0f557b15892c29ab5beabbc51aa8dc791239b70cb"} Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.117056 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.348835 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" podStartSLOduration=3.348815793 podStartE2EDuration="3.348815793s" podCreationTimestamp="2025-11-28 11:42:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:42:29.160068032 +0000 UTC m=+1226.692581953" watchObservedRunningTime="2025-11-28 11:42:29.348815793 +0000 UTC m=+1226.881329714" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.433475 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-5pnwp"] Nov 28 11:42:29 crc kubenswrapper[4862]: E1128 11:42:29.433972 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51e2e8d-a77a-4017-90bd-1e2dfdc8f527" containerName="mariadb-database-create" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.433997 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51e2e8d-a77a-4017-90bd-1e2dfdc8f527" containerName="mariadb-database-create" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.434212 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51e2e8d-a77a-4017-90bd-1e2dfdc8f527" containerName="mariadb-database-create" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.434924 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.442415 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.443618 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-056c-account-create-update-4t96j"] Nov 28 11:42:29 crc kubenswrapper[4862]: E1128 11:42:29.444189 4862 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 11:42:29 crc kubenswrapper[4862]: E1128 11:42:29.444219 4862 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 11:42:29 crc kubenswrapper[4862]: E1128 11:42:29.444274 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift podName:b56372b7-fc39-40cc-8e0a-e95f907db697 nodeName:}" failed. No retries permitted until 2025-11-28 11:42:31.444254127 +0000 UTC m=+1228.976768048 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift") pod "swift-storage-0" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697") : configmap "swift-ring-files" not found Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.445983 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.449836 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.466770 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5pnwp"] Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.480083 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-056c-account-create-update-4t96j"] Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.553326 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzkkp\" (UniqueName: \"kubernetes.io/projected/0e88fe61-1181-4385-8d0a-d99483e8112f-kube-api-access-mzkkp\") pod \"glance-db-create-5pnwp\" (UID: \"0e88fe61-1181-4385-8d0a-d99483e8112f\") " pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.553406 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8kll\" (UniqueName: \"kubernetes.io/projected/ff113c73-a3c6-4300-93f2-fed5c3a1040f-kube-api-access-k8kll\") pod \"glance-056c-account-create-update-4t96j\" (UID: \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\") " pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.553527 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e88fe61-1181-4385-8d0a-d99483e8112f-operator-scripts\") pod \"glance-db-create-5pnwp\" (UID: \"0e88fe61-1181-4385-8d0a-d99483e8112f\") " pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.553581 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff113c73-a3c6-4300-93f2-fed5c3a1040f-operator-scripts\") pod \"glance-056c-account-create-update-4t96j\" (UID: \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\") " pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.643489 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.655086 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff113c73-a3c6-4300-93f2-fed5c3a1040f-operator-scripts\") pod \"glance-056c-account-create-update-4t96j\" (UID: \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\") " pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.655270 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzkkp\" (UniqueName: \"kubernetes.io/projected/0e88fe61-1181-4385-8d0a-d99483e8112f-kube-api-access-mzkkp\") pod \"glance-db-create-5pnwp\" (UID: \"0e88fe61-1181-4385-8d0a-d99483e8112f\") " pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.655420 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8kll\" (UniqueName: \"kubernetes.io/projected/ff113c73-a3c6-4300-93f2-fed5c3a1040f-kube-api-access-k8kll\") pod \"glance-056c-account-create-update-4t96j\" (UID: \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\") " pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.655614 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e88fe61-1181-4385-8d0a-d99483e8112f-operator-scripts\") pod \"glance-db-create-5pnwp\" (UID: \"0e88fe61-1181-4385-8d0a-d99483e8112f\") " pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.655763 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff113c73-a3c6-4300-93f2-fed5c3a1040f-operator-scripts\") pod \"glance-056c-account-create-update-4t96j\" (UID: \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\") " pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.656495 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e88fe61-1181-4385-8d0a-d99483e8112f-operator-scripts\") pod \"glance-db-create-5pnwp\" (UID: \"0e88fe61-1181-4385-8d0a-d99483e8112f\") " pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.679080 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzkkp\" (UniqueName: \"kubernetes.io/projected/0e88fe61-1181-4385-8d0a-d99483e8112f-kube-api-access-mzkkp\") pod \"glance-db-create-5pnwp\" (UID: \"0e88fe61-1181-4385-8d0a-d99483e8112f\") " pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.679668 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8kll\" (UniqueName: \"kubernetes.io/projected/ff113c73-a3c6-4300-93f2-fed5c3a1040f-kube-api-access-k8kll\") pod \"glance-056c-account-create-update-4t96j\" (UID: \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\") " pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.757568 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fhtb\" (UniqueName: \"kubernetes.io/projected/705faed3-031b-4ec7-bfd7-b5ac626c98b3-kube-api-access-4fhtb\") pod \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\" (UID: \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\") " Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.757819 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/705faed3-031b-4ec7-bfd7-b5ac626c98b3-operator-scripts\") pod \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\" (UID: \"705faed3-031b-4ec7-bfd7-b5ac626c98b3\") " Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.758551 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/705faed3-031b-4ec7-bfd7-b5ac626c98b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "705faed3-031b-4ec7-bfd7-b5ac626c98b3" (UID: "705faed3-031b-4ec7-bfd7-b5ac626c98b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.759985 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.760742 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/705faed3-031b-4ec7-bfd7-b5ac626c98b3-kube-api-access-4fhtb" (OuterVolumeSpecName: "kube-api-access-4fhtb") pod "705faed3-031b-4ec7-bfd7-b5ac626c98b3" (UID: "705faed3-031b-4ec7-bfd7-b5ac626c98b3"). InnerVolumeSpecName "kube-api-access-4fhtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.775550 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.857396 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.863001 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/705faed3-031b-4ec7-bfd7-b5ac626c98b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.863176 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fhtb\" (UniqueName: \"kubernetes.io/projected/705faed3-031b-4ec7-bfd7-b5ac626c98b3-kube-api-access-4fhtb\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.863195 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.964820 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b89v6\" (UniqueName: \"kubernetes.io/projected/1ec7a6dc-a28c-4239-b112-6139ae089cee-kube-api-access-b89v6\") pod \"1ec7a6dc-a28c-4239-b112-6139ae089cee\" (UID: \"1ec7a6dc-a28c-4239-b112-6139ae089cee\") " Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.965182 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxwvr\" (UniqueName: \"kubernetes.io/projected/d990f484-bffa-4bfc-8c45-b00c5514512a-kube-api-access-sxwvr\") pod \"d990f484-bffa-4bfc-8c45-b00c5514512a\" (UID: \"d990f484-bffa-4bfc-8c45-b00c5514512a\") " Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.965232 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec7a6dc-a28c-4239-b112-6139ae089cee-operator-scripts\") pod \"1ec7a6dc-a28c-4239-b112-6139ae089cee\" (UID: \"1ec7a6dc-a28c-4239-b112-6139ae089cee\") " Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.965269 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d990f484-bffa-4bfc-8c45-b00c5514512a-operator-scripts\") pod \"d990f484-bffa-4bfc-8c45-b00c5514512a\" (UID: \"d990f484-bffa-4bfc-8c45-b00c5514512a\") " Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.965928 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec7a6dc-a28c-4239-b112-6139ae089cee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ec7a6dc-a28c-4239-b112-6139ae089cee" (UID: "1ec7a6dc-a28c-4239-b112-6139ae089cee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.966476 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d990f484-bffa-4bfc-8c45-b00c5514512a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d990f484-bffa-4bfc-8c45-b00c5514512a" (UID: "d990f484-bffa-4bfc-8c45-b00c5514512a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.971327 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec7a6dc-a28c-4239-b112-6139ae089cee-kube-api-access-b89v6" (OuterVolumeSpecName: "kube-api-access-b89v6") pod "1ec7a6dc-a28c-4239-b112-6139ae089cee" (UID: "1ec7a6dc-a28c-4239-b112-6139ae089cee"). InnerVolumeSpecName "kube-api-access-b89v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:29 crc kubenswrapper[4862]: I1128 11:42:29.973353 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d990f484-bffa-4bfc-8c45-b00c5514512a-kube-api-access-sxwvr" (OuterVolumeSpecName: "kube-api-access-sxwvr") pod "d990f484-bffa-4bfc-8c45-b00c5514512a" (UID: "d990f484-bffa-4bfc-8c45-b00c5514512a"). InnerVolumeSpecName "kube-api-access-sxwvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.067732 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b89v6\" (UniqueName: \"kubernetes.io/projected/1ec7a6dc-a28c-4239-b112-6139ae089cee-kube-api-access-b89v6\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.067764 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxwvr\" (UniqueName: \"kubernetes.io/projected/d990f484-bffa-4bfc-8c45-b00c5514512a-kube-api-access-sxwvr\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.067774 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec7a6dc-a28c-4239-b112-6139ae089cee-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.067783 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d990f484-bffa-4bfc-8c45-b00c5514512a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.126033 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q4jzk" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.126743 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q4jzk" event={"ID":"1ec7a6dc-a28c-4239-b112-6139ae089cee","Type":"ContainerDied","Data":"784fefb4c2627b19d12af7de42886cb1b9f414241d8468f68a9bc666c0d14a1f"} Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.126789 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="784fefb4c2627b19d12af7de42886cb1b9f414241d8468f68a9bc666c0d14a1f" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.128449 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bc3a-account-create-update-x9czv" event={"ID":"705faed3-031b-4ec7-bfd7-b5ac626c98b3","Type":"ContainerDied","Data":"71891aebfed6dbf2274befd359365a530000a88a41286ec99c76ef3fbb1ec650"} Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.128465 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71891aebfed6dbf2274befd359365a530000a88a41286ec99c76ef3fbb1ec650" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.128518 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bc3a-account-create-update-x9czv" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.139177 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e4ca-account-create-update-hsh77" event={"ID":"d990f484-bffa-4bfc-8c45-b00c5514512a","Type":"ContainerDied","Data":"6ced57fa1cbe12c8075eb1964a75a704c9c38e6f6af5b27dceab9c29c11c88f5"} Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.139210 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ced57fa1cbe12c8075eb1964a75a704c9c38e6f6af5b27dceab9c29c11c88f5" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.139226 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e4ca-account-create-update-hsh77" Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.332463 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-056c-account-create-update-4t96j"] Nov 28 11:42:30 crc kubenswrapper[4862]: W1128 11:42:30.344420 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff113c73_a3c6_4300_93f2_fed5c3a1040f.slice/crio-92d025efa70c425e98a19704081be7777a91de5e40bcd6b9e2665f2cd8cad02c WatchSource:0}: Error finding container 92d025efa70c425e98a19704081be7777a91de5e40bcd6b9e2665f2cd8cad02c: Status 404 returned error can't find the container with id 92d025efa70c425e98a19704081be7777a91de5e40bcd6b9e2665f2cd8cad02c Nov 28 11:42:30 crc kubenswrapper[4862]: W1128 11:42:30.349808 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e88fe61_1181_4385_8d0a_d99483e8112f.slice/crio-ce973bcd9967b0cee8b0ed1ff190da28ff7c2f0a3b336c7f430a85226323f6c6 WatchSource:0}: Error finding container ce973bcd9967b0cee8b0ed1ff190da28ff7c2f0a3b336c7f430a85226323f6c6: Status 404 returned error can't find the container with id ce973bcd9967b0cee8b0ed1ff190da28ff7c2f0a3b336c7f430a85226323f6c6 Nov 28 11:42:30 crc kubenswrapper[4862]: I1128 11:42:30.350633 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5pnwp"] Nov 28 11:42:31 crc kubenswrapper[4862]: I1128 11:42:31.158616 4862 generic.go:334] "Generic (PLEG): container finished" podID="ff113c73-a3c6-4300-93f2-fed5c3a1040f" containerID="12cc858922f84aa30fabdabce429fbe3a12d1c5905d381e7d6b7a6811417f1a1" exitCode=0 Nov 28 11:42:31 crc kubenswrapper[4862]: I1128 11:42:31.158829 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-056c-account-create-update-4t96j" event={"ID":"ff113c73-a3c6-4300-93f2-fed5c3a1040f","Type":"ContainerDied","Data":"12cc858922f84aa30fabdabce429fbe3a12d1c5905d381e7d6b7a6811417f1a1"} Nov 28 11:42:31 crc kubenswrapper[4862]: I1128 11:42:31.159182 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-056c-account-create-update-4t96j" event={"ID":"ff113c73-a3c6-4300-93f2-fed5c3a1040f","Type":"ContainerStarted","Data":"92d025efa70c425e98a19704081be7777a91de5e40bcd6b9e2665f2cd8cad02c"} Nov 28 11:42:31 crc kubenswrapper[4862]: I1128 11:42:31.165002 4862 generic.go:334] "Generic (PLEG): container finished" podID="0e88fe61-1181-4385-8d0a-d99483e8112f" containerID="b6aaf41414c9fd41c763d2add37d020134b5bae524760cad7b8d323eb5a18a35" exitCode=0 Nov 28 11:42:31 crc kubenswrapper[4862]: I1128 11:42:31.166339 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5pnwp" event={"ID":"0e88fe61-1181-4385-8d0a-d99483e8112f","Type":"ContainerDied","Data":"b6aaf41414c9fd41c763d2add37d020134b5bae524760cad7b8d323eb5a18a35"} Nov 28 11:42:31 crc kubenswrapper[4862]: I1128 11:42:31.167314 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5pnwp" event={"ID":"0e88fe61-1181-4385-8d0a-d99483e8112f","Type":"ContainerStarted","Data":"ce973bcd9967b0cee8b0ed1ff190da28ff7c2f0a3b336c7f430a85226323f6c6"} Nov 28 11:42:31 crc kubenswrapper[4862]: I1128 11:42:31.493641 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:31 crc kubenswrapper[4862]: E1128 11:42:31.493825 4862 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 11:42:31 crc kubenswrapper[4862]: E1128 11:42:31.493848 4862 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 11:42:31 crc kubenswrapper[4862]: E1128 11:42:31.493905 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift podName:b56372b7-fc39-40cc-8e0a-e95f907db697 nodeName:}" failed. No retries permitted until 2025-11-28 11:42:35.493888095 +0000 UTC m=+1233.026402006 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift") pod "swift-storage-0" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697") : configmap "swift-ring-files" not found Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.574485 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.581451 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.633775 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8kll\" (UniqueName: \"kubernetes.io/projected/ff113c73-a3c6-4300-93f2-fed5c3a1040f-kube-api-access-k8kll\") pod \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\" (UID: \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\") " Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.634145 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzkkp\" (UniqueName: \"kubernetes.io/projected/0e88fe61-1181-4385-8d0a-d99483e8112f-kube-api-access-mzkkp\") pod \"0e88fe61-1181-4385-8d0a-d99483e8112f\" (UID: \"0e88fe61-1181-4385-8d0a-d99483e8112f\") " Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.634209 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e88fe61-1181-4385-8d0a-d99483e8112f-operator-scripts\") pod \"0e88fe61-1181-4385-8d0a-d99483e8112f\" (UID: \"0e88fe61-1181-4385-8d0a-d99483e8112f\") " Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.634361 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff113c73-a3c6-4300-93f2-fed5c3a1040f-operator-scripts\") pod \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\" (UID: \"ff113c73-a3c6-4300-93f2-fed5c3a1040f\") " Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.635832 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff113c73-a3c6-4300-93f2-fed5c3a1040f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff113c73-a3c6-4300-93f2-fed5c3a1040f" (UID: "ff113c73-a3c6-4300-93f2-fed5c3a1040f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.640212 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff113c73-a3c6-4300-93f2-fed5c3a1040f-kube-api-access-k8kll" (OuterVolumeSpecName: "kube-api-access-k8kll") pod "ff113c73-a3c6-4300-93f2-fed5c3a1040f" (UID: "ff113c73-a3c6-4300-93f2-fed5c3a1040f"). InnerVolumeSpecName "kube-api-access-k8kll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.641816 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e88fe61-1181-4385-8d0a-d99483e8112f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e88fe61-1181-4385-8d0a-d99483e8112f" (UID: "0e88fe61-1181-4385-8d0a-d99483e8112f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.645487 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e88fe61-1181-4385-8d0a-d99483e8112f-kube-api-access-mzkkp" (OuterVolumeSpecName: "kube-api-access-mzkkp") pod "0e88fe61-1181-4385-8d0a-d99483e8112f" (UID: "0e88fe61-1181-4385-8d0a-d99483e8112f"). InnerVolumeSpecName "kube-api-access-mzkkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.736506 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff113c73-a3c6-4300-93f2-fed5c3a1040f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.736540 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8kll\" (UniqueName: \"kubernetes.io/projected/ff113c73-a3c6-4300-93f2-fed5c3a1040f-kube-api-access-k8kll\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.736556 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzkkp\" (UniqueName: \"kubernetes.io/projected/0e88fe61-1181-4385-8d0a-d99483e8112f-kube-api-access-mzkkp\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.736567 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e88fe61-1181-4385-8d0a-d99483e8112f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:33 crc kubenswrapper[4862]: I1128 11:42:33.771291 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 28 11:42:34 crc kubenswrapper[4862]: I1128 11:42:34.195536 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xchsv" event={"ID":"3ac70dca-3c3a-43fe-81ef-2677c476e9a4","Type":"ContainerStarted","Data":"11f076f97304361292994e05f8355c71c3532228c8bc7325739454e7865993d0"} Nov 28 11:42:34 crc kubenswrapper[4862]: I1128 11:42:34.218837 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-056c-account-create-update-4t96j" event={"ID":"ff113c73-a3c6-4300-93f2-fed5c3a1040f","Type":"ContainerDied","Data":"92d025efa70c425e98a19704081be7777a91de5e40bcd6b9e2665f2cd8cad02c"} Nov 28 11:42:34 crc kubenswrapper[4862]: I1128 11:42:34.218889 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92d025efa70c425e98a19704081be7777a91de5e40bcd6b9e2665f2cd8cad02c" Nov 28 11:42:34 crc kubenswrapper[4862]: I1128 11:42:34.219045 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-056c-account-create-update-4t96j" Nov 28 11:42:34 crc kubenswrapper[4862]: I1128 11:42:34.220930 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5pnwp" event={"ID":"0e88fe61-1181-4385-8d0a-d99483e8112f","Type":"ContainerDied","Data":"ce973bcd9967b0cee8b0ed1ff190da28ff7c2f0a3b336c7f430a85226323f6c6"} Nov 28 11:42:34 crc kubenswrapper[4862]: I1128 11:42:34.221016 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce973bcd9967b0cee8b0ed1ff190da28ff7c2f0a3b336c7f430a85226323f6c6" Nov 28 11:42:34 crc kubenswrapper[4862]: I1128 11:42:34.221131 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5pnwp" Nov 28 11:42:34 crc kubenswrapper[4862]: I1128 11:42:34.248296 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-xchsv" podStartSLOduration=1.653323697 podStartE2EDuration="6.248278864s" podCreationTimestamp="2025-11-28 11:42:28 +0000 UTC" firstStartedPulling="2025-11-28 11:42:28.814503287 +0000 UTC m=+1226.347017208" lastFinishedPulling="2025-11-28 11:42:33.409458454 +0000 UTC m=+1230.941972375" observedRunningTime="2025-11-28 11:42:34.243628841 +0000 UTC m=+1231.776142762" watchObservedRunningTime="2025-11-28 11:42:34.248278864 +0000 UTC m=+1231.780792785" Nov 28 11:42:35 crc kubenswrapper[4862]: I1128 11:42:35.576157 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:35 crc kubenswrapper[4862]: E1128 11:42:35.576777 4862 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 11:42:35 crc kubenswrapper[4862]: E1128 11:42:35.576805 4862 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 11:42:35 crc kubenswrapper[4862]: E1128 11:42:35.576880 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift podName:b56372b7-fc39-40cc-8e0a-e95f907db697 nodeName:}" failed. No retries permitted until 2025-11-28 11:42:43.576850568 +0000 UTC m=+1241.109364529 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift") pod "swift-storage-0" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697") : configmap "swift-ring-files" not found Nov 28 11:42:36 crc kubenswrapper[4862]: I1128 11:42:36.969341 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:42:37 crc kubenswrapper[4862]: I1128 11:42:37.051519 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-c6fvq"] Nov 28 11:42:37 crc kubenswrapper[4862]: I1128 11:42:37.052365 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" podUID="fa0af7e6-4966-4915-9807-d2055f09c111" containerName="dnsmasq-dns" containerID="cri-o://ee0d80ae8fee71d34c192ca3acdb1458773fb691afab6adc3870c00b988311d1" gracePeriod=10 Nov 28 11:42:38 crc kubenswrapper[4862]: I1128 11:42:38.291401 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:42:38 crc kubenswrapper[4862]: I1128 11:42:38.292179 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:42:38 crc kubenswrapper[4862]: I1128 11:42:38.292264 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:42:38 crc kubenswrapper[4862]: I1128 11:42:38.293512 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1439dec9a32df00b0169c77b7376186b3da92973a1be5f562a161099f8fb96c8"} pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:42:38 crc kubenswrapper[4862]: I1128 11:42:38.293628 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" containerID="cri-o://1439dec9a32df00b0169c77b7376186b3da92973a1be5f562a161099f8fb96c8" gracePeriod=600 Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.278703 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6845439-1a4c-4f28-9ba4-29149d545995" containerID="1439dec9a32df00b0169c77b7376186b3da92973a1be5f562a161099f8fb96c8" exitCode=0 Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.278821 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerDied","Data":"1439dec9a32df00b0169c77b7376186b3da92973a1be5f562a161099f8fb96c8"} Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.279128 4862 scope.go:117] "RemoveContainer" containerID="5e63a1b14a90646d7f5b8ec0b596a3abf9ade9420df0f277e56e83cc3a51ca76" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.282029 4862 generic.go:334] "Generic (PLEG): container finished" podID="fa0af7e6-4966-4915-9807-d2055f09c111" containerID="ee0d80ae8fee71d34c192ca3acdb1458773fb691afab6adc3870c00b988311d1" exitCode=0 Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.282064 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" event={"ID":"fa0af7e6-4966-4915-9807-d2055f09c111","Type":"ContainerDied","Data":"ee0d80ae8fee71d34c192ca3acdb1458773fb691afab6adc3870c00b988311d1"} Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.407304 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.553909 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-dns-svc\") pod \"fa0af7e6-4966-4915-9807-d2055f09c111\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.553987 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4w82\" (UniqueName: \"kubernetes.io/projected/fa0af7e6-4966-4915-9807-d2055f09c111-kube-api-access-m4w82\") pod \"fa0af7e6-4966-4915-9807-d2055f09c111\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.554014 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-config\") pod \"fa0af7e6-4966-4915-9807-d2055f09c111\" (UID: \"fa0af7e6-4966-4915-9807-d2055f09c111\") " Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.560300 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa0af7e6-4966-4915-9807-d2055f09c111-kube-api-access-m4w82" (OuterVolumeSpecName: "kube-api-access-m4w82") pod "fa0af7e6-4966-4915-9807-d2055f09c111" (UID: "fa0af7e6-4966-4915-9807-d2055f09c111"). InnerVolumeSpecName "kube-api-access-m4w82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.606167 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-qltsz"] Nov 28 11:42:39 crc kubenswrapper[4862]: E1128 11:42:39.607013 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff113c73-a3c6-4300-93f2-fed5c3a1040f" containerName="mariadb-account-create-update" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.607180 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff113c73-a3c6-4300-93f2-fed5c3a1040f" containerName="mariadb-account-create-update" Nov 28 11:42:39 crc kubenswrapper[4862]: E1128 11:42:39.607332 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d990f484-bffa-4bfc-8c45-b00c5514512a" containerName="mariadb-account-create-update" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.607451 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="d990f484-bffa-4bfc-8c45-b00c5514512a" containerName="mariadb-account-create-update" Nov 28 11:42:39 crc kubenswrapper[4862]: E1128 11:42:39.607570 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="705faed3-031b-4ec7-bfd7-b5ac626c98b3" containerName="mariadb-account-create-update" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.607672 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="705faed3-031b-4ec7-bfd7-b5ac626c98b3" containerName="mariadb-account-create-update" Nov 28 11:42:39 crc kubenswrapper[4862]: E1128 11:42:39.607842 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0af7e6-4966-4915-9807-d2055f09c111" containerName="init" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.607963 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0af7e6-4966-4915-9807-d2055f09c111" containerName="init" Nov 28 11:42:39 crc kubenswrapper[4862]: E1128 11:42:39.608133 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e88fe61-1181-4385-8d0a-d99483e8112f" containerName="mariadb-database-create" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.608244 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e88fe61-1181-4385-8d0a-d99483e8112f" containerName="mariadb-database-create" Nov 28 11:42:39 crc kubenswrapper[4862]: E1128 11:42:39.608357 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec7a6dc-a28c-4239-b112-6139ae089cee" containerName="mariadb-database-create" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.608460 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec7a6dc-a28c-4239-b112-6139ae089cee" containerName="mariadb-database-create" Nov 28 11:42:39 crc kubenswrapper[4862]: E1128 11:42:39.608582 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0af7e6-4966-4915-9807-d2055f09c111" containerName="dnsmasq-dns" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.608697 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0af7e6-4966-4915-9807-d2055f09c111" containerName="dnsmasq-dns" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.609125 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="d990f484-bffa-4bfc-8c45-b00c5514512a" containerName="mariadb-account-create-update" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.609276 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e88fe61-1181-4385-8d0a-d99483e8112f" containerName="mariadb-database-create" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.609399 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0af7e6-4966-4915-9807-d2055f09c111" containerName="dnsmasq-dns" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.609506 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ec7a6dc-a28c-4239-b112-6139ae089cee" containerName="mariadb-database-create" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.609633 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="705faed3-031b-4ec7-bfd7-b5ac626c98b3" containerName="mariadb-account-create-update" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.609762 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff113c73-a3c6-4300-93f2-fed5c3a1040f" containerName="mariadb-account-create-update" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.610770 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.613720 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.614141 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vp6qr" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.630302 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qltsz"] Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.639176 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-config" (OuterVolumeSpecName: "config") pod "fa0af7e6-4966-4915-9807-d2055f09c111" (UID: "fa0af7e6-4966-4915-9807-d2055f09c111"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.650539 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fa0af7e6-4966-4915-9807-d2055f09c111" (UID: "fa0af7e6-4966-4915-9807-d2055f09c111"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.656280 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.656347 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4w82\" (UniqueName: \"kubernetes.io/projected/fa0af7e6-4966-4915-9807-d2055f09c111-kube-api-access-m4w82\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.656436 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa0af7e6-4966-4915-9807-d2055f09c111-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.758294 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-config-data\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.758380 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-db-sync-config-data\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.758435 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vfpp\" (UniqueName: \"kubernetes.io/projected/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-kube-api-access-6vfpp\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.758479 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-combined-ca-bundle\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.859596 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-config-data\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.859678 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-db-sync-config-data\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.859736 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vfpp\" (UniqueName: \"kubernetes.io/projected/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-kube-api-access-6vfpp\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.859779 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-combined-ca-bundle\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.865485 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-db-sync-config-data\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.866012 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-config-data\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.868453 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-combined-ca-bundle\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:39 crc kubenswrapper[4862]: I1128 11:42:39.885268 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vfpp\" (UniqueName: \"kubernetes.io/projected/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-kube-api-access-6vfpp\") pod \"glance-db-sync-qltsz\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.035683 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qltsz" Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.293495 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"9f1769f4e9e7625f7b202c55e00f4690f331fdd2c1054c72cfd81ec88107ff67"} Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.296742 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" event={"ID":"fa0af7e6-4966-4915-9807-d2055f09c111","Type":"ContainerDied","Data":"2bb8690f3555c12056e09baca4783a483520c623094d98344166f0f1747ee4d7"} Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.296787 4862 scope.go:117] "RemoveContainer" containerID="ee0d80ae8fee71d34c192ca3acdb1458773fb691afab6adc3870c00b988311d1" Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.296876 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-c6fvq" Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.325609 4862 scope.go:117] "RemoveContainer" containerID="76c619f1dbd52eb0e8a5b8c88d7dd5661d3fdec55812c8feb8f4fca1b5a9d61f" Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.338481 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-c6fvq"] Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.349051 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-c6fvq"] Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.683818 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qltsz"] Nov 28 11:42:40 crc kubenswrapper[4862]: W1128 11:42:40.687012 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod010a34d3_73a3_4f8e_9bc3_15bb24d1d41f.slice/crio-efe828311882e481520b14baa1aa5981275605b518a440e5a64a12977ad00b70 WatchSource:0}: Error finding container efe828311882e481520b14baa1aa5981275605b518a440e5a64a12977ad00b70: Status 404 returned error can't find the container with id efe828311882e481520b14baa1aa5981275605b518a440e5a64a12977ad00b70 Nov 28 11:42:40 crc kubenswrapper[4862]: I1128 11:42:40.850205 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa0af7e6-4966-4915-9807-d2055f09c111" path="/var/lib/kubelet/pods/fa0af7e6-4966-4915-9807-d2055f09c111/volumes" Nov 28 11:42:41 crc kubenswrapper[4862]: I1128 11:42:41.307203 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qltsz" event={"ID":"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f","Type":"ContainerStarted","Data":"efe828311882e481520b14baa1aa5981275605b518a440e5a64a12977ad00b70"} Nov 28 11:42:43 crc kubenswrapper[4862]: I1128 11:42:43.626260 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:43 crc kubenswrapper[4862]: E1128 11:42:43.626478 4862 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 11:42:43 crc kubenswrapper[4862]: E1128 11:42:43.626984 4862 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 11:42:43 crc kubenswrapper[4862]: E1128 11:42:43.627083 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift podName:b56372b7-fc39-40cc-8e0a-e95f907db697 nodeName:}" failed. No retries permitted until 2025-11-28 11:42:59.627048003 +0000 UTC m=+1257.159561954 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift") pod "swift-storage-0" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697") : configmap "swift-ring-files" not found Nov 28 11:42:44 crc kubenswrapper[4862]: I1128 11:42:44.338488 4862 generic.go:334] "Generic (PLEG): container finished" podID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" containerID="28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245" exitCode=0 Nov 28 11:42:44 crc kubenswrapper[4862]: I1128 11:42:44.338564 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356","Type":"ContainerDied","Data":"28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245"} Nov 28 11:42:44 crc kubenswrapper[4862]: I1128 11:42:44.343624 4862 generic.go:334] "Generic (PLEG): container finished" podID="3ac70dca-3c3a-43fe-81ef-2677c476e9a4" containerID="11f076f97304361292994e05f8355c71c3532228c8bc7325739454e7865993d0" exitCode=0 Nov 28 11:42:44 crc kubenswrapper[4862]: I1128 11:42:44.343694 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xchsv" event={"ID":"3ac70dca-3c3a-43fe-81ef-2677c476e9a4","Type":"ContainerDied","Data":"11f076f97304361292994e05f8355c71c3532228c8bc7325739454e7865993d0"} Nov 28 11:42:44 crc kubenswrapper[4862]: I1128 11:42:44.969771 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:42:44 crc kubenswrapper[4862]: I1128 11:42:44.978794 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:42:44 crc kubenswrapper[4862]: I1128 11:42:44.979665 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hkgwr" podUID="b1bdf176-776a-4211-8b43-ff90225de25a" containerName="ovn-controller" probeResult="failure" output=< Nov 28 11:42:44 crc kubenswrapper[4862]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 28 11:42:44 crc kubenswrapper[4862]: > Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.198837 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hkgwr-config-jzcz8"] Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.200707 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.202473 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.239742 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hkgwr-config-jzcz8"] Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.355904 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356","Type":"ContainerStarted","Data":"dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f"} Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.356423 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.364375 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run-ovn\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.364454 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-scripts\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.364579 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m87wm\" (UniqueName: \"kubernetes.io/projected/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-kube-api-access-m87wm\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.364622 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-additional-scripts\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.364673 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-log-ovn\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.364697 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.384681 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=58.570069539 podStartE2EDuration="1m6.384656728s" podCreationTimestamp="2025-11-28 11:41:39 +0000 UTC" firstStartedPulling="2025-11-28 11:42:01.120518458 +0000 UTC m=+1198.653032379" lastFinishedPulling="2025-11-28 11:42:08.935105647 +0000 UTC m=+1206.467619568" observedRunningTime="2025-11-28 11:42:45.378506154 +0000 UTC m=+1242.911020065" watchObservedRunningTime="2025-11-28 11:42:45.384656728 +0000 UTC m=+1242.917170649" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.466257 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run-ovn\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.466300 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-scripts\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.466411 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m87wm\" (UniqueName: \"kubernetes.io/projected/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-kube-api-access-m87wm\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.466449 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-additional-scripts\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.466470 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-log-ovn\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.466488 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.466595 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run-ovn\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.467329 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-log-ovn\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.467374 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.468022 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-additional-scripts\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.469864 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-scripts\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.515499 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m87wm\" (UniqueName: \"kubernetes.io/projected/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-kube-api-access-m87wm\") pod \"ovn-controller-hkgwr-config-jzcz8\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:45 crc kubenswrapper[4862]: I1128 11:42:45.554444 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:49 crc kubenswrapper[4862]: I1128 11:42:49.980921 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hkgwr" podUID="b1bdf176-776a-4211-8b43-ff90225de25a" containerName="ovn-controller" probeResult="failure" output=< Nov 28 11:42:49 crc kubenswrapper[4862]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 28 11:42:49 crc kubenswrapper[4862]: > Nov 28 11:42:50 crc kubenswrapper[4862]: I1128 11:42:50.419557 4862 generic.go:334] "Generic (PLEG): container finished" podID="527b2bce-1186-4186-8992-a3a63d2d2b22" containerID="b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2" exitCode=0 Nov 28 11:42:50 crc kubenswrapper[4862]: I1128 11:42:50.419596 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"527b2bce-1186-4186-8992-a3a63d2d2b22","Type":"ContainerDied","Data":"b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2"} Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.337441 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.426565 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-scripts\") pod \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.426702 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-combined-ca-bundle\") pod \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.426771 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kklvf\" (UniqueName: \"kubernetes.io/projected/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-kube-api-access-kklvf\") pod \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.426803 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-etc-swift\") pod \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.426916 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-ring-data-devices\") pod \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.427010 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-swiftconf\") pod \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.427045 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-dispersionconf\") pod \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\" (UID: \"3ac70dca-3c3a-43fe-81ef-2677c476e9a4\") " Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.427787 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3ac70dca-3c3a-43fe-81ef-2677c476e9a4" (UID: "3ac70dca-3c3a-43fe-81ef-2677c476e9a4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.428567 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3ac70dca-3c3a-43fe-81ef-2677c476e9a4" (UID: "3ac70dca-3c3a-43fe-81ef-2677c476e9a4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.433189 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-kube-api-access-kklvf" (OuterVolumeSpecName: "kube-api-access-kklvf") pod "3ac70dca-3c3a-43fe-81ef-2677c476e9a4" (UID: "3ac70dca-3c3a-43fe-81ef-2677c476e9a4"). InnerVolumeSpecName "kube-api-access-kklvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.435386 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3ac70dca-3c3a-43fe-81ef-2677c476e9a4" (UID: "3ac70dca-3c3a-43fe-81ef-2677c476e9a4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.453411 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ac70dca-3c3a-43fe-81ef-2677c476e9a4" (UID: "3ac70dca-3c3a-43fe-81ef-2677c476e9a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.458605 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3ac70dca-3c3a-43fe-81ef-2677c476e9a4" (UID: "3ac70dca-3c3a-43fe-81ef-2677c476e9a4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.459269 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-scripts" (OuterVolumeSpecName: "scripts") pod "3ac70dca-3c3a-43fe-81ef-2677c476e9a4" (UID: "3ac70dca-3c3a-43fe-81ef-2677c476e9a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.461995 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xchsv" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.462008 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xchsv" event={"ID":"3ac70dca-3c3a-43fe-81ef-2677c476e9a4","Type":"ContainerDied","Data":"21261a8d996fd1db9b549697ba9c74193b4ad2804231cbb8e9f0cee7e36fb3ad"} Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.462296 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21261a8d996fd1db9b549697ba9c74193b4ad2804231cbb8e9f0cee7e36fb3ad" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.464050 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"527b2bce-1186-4186-8992-a3a63d2d2b22","Type":"ContainerStarted","Data":"d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97"} Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.464616 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.494125 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=64.67635217 podStartE2EDuration="1m13.494108484s" podCreationTimestamp="2025-11-28 11:41:40 +0000 UTC" firstStartedPulling="2025-11-28 11:42:00.782667327 +0000 UTC m=+1198.315181248" lastFinishedPulling="2025-11-28 11:42:09.600423641 +0000 UTC m=+1207.132937562" observedRunningTime="2025-11-28 11:42:53.49019536 +0000 UTC m=+1251.022709281" watchObservedRunningTime="2025-11-28 11:42:53.494108484 +0000 UTC m=+1251.026622405" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.529857 4862 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.529884 4862 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.529892 4862 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.529902 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.529910 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.529919 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kklvf\" (UniqueName: \"kubernetes.io/projected/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-kube-api-access-kklvf\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.529928 4862 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ac70dca-3c3a-43fe-81ef-2677c476e9a4-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:53 crc kubenswrapper[4862]: I1128 11:42:53.643815 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hkgwr-config-jzcz8"] Nov 28 11:42:53 crc kubenswrapper[4862]: W1128 11:42:53.647074 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6cfbfd1_9d5e_49b8_99ce_b6fd04ae077c.slice/crio-5aa1d7841d9ffd14387be208cec6bd9f3daeef0b7c08dc4125cd4d06615a3b25 WatchSource:0}: Error finding container 5aa1d7841d9ffd14387be208cec6bd9f3daeef0b7c08dc4125cd4d06615a3b25: Status 404 returned error can't find the container with id 5aa1d7841d9ffd14387be208cec6bd9f3daeef0b7c08dc4125cd4d06615a3b25 Nov 28 11:42:54 crc kubenswrapper[4862]: I1128 11:42:54.474229 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" containerID="fd35b665be415af27b736bab6a7dda350802c5432b416ed0360895ab52239829" exitCode=0 Nov 28 11:42:54 crc kubenswrapper[4862]: I1128 11:42:54.474331 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hkgwr-config-jzcz8" event={"ID":"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c","Type":"ContainerDied","Data":"fd35b665be415af27b736bab6a7dda350802c5432b416ed0360895ab52239829"} Nov 28 11:42:54 crc kubenswrapper[4862]: I1128 11:42:54.474878 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hkgwr-config-jzcz8" event={"ID":"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c","Type":"ContainerStarted","Data":"5aa1d7841d9ffd14387be208cec6bd9f3daeef0b7c08dc4125cd4d06615a3b25"} Nov 28 11:42:54 crc kubenswrapper[4862]: I1128 11:42:54.476708 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qltsz" event={"ID":"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f","Type":"ContainerStarted","Data":"00badb80af0f696931c741191dda1d4b64ccf8e89732147e270eb8afbaa447c3"} Nov 28 11:42:54 crc kubenswrapper[4862]: I1128 11:42:54.509429 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-qltsz" podStartSLOduration=2.924628974 podStartE2EDuration="15.509405341s" podCreationTimestamp="2025-11-28 11:42:39 +0000 UTC" firstStartedPulling="2025-11-28 11:42:40.689659635 +0000 UTC m=+1238.222173566" lastFinishedPulling="2025-11-28 11:42:53.274436012 +0000 UTC m=+1250.806949933" observedRunningTime="2025-11-28 11:42:54.504520751 +0000 UTC m=+1252.037034672" watchObservedRunningTime="2025-11-28 11:42:54.509405341 +0000 UTC m=+1252.041919262" Nov 28 11:42:54 crc kubenswrapper[4862]: I1128 11:42:54.964034 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-hkgwr" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.829720 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.975872 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run-ovn\") pod \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.976193 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-additional-scripts\") pod \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.976347 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m87wm\" (UniqueName: \"kubernetes.io/projected/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-kube-api-access-m87wm\") pod \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.976513 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-log-ovn\") pod \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.976703 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-scripts\") pod \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.976834 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run\") pod \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\" (UID: \"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c\") " Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.975974 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" (UID: "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.976566 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" (UID: "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.976923 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" (UID: "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.977431 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run" (OuterVolumeSpecName: "var-run") pod "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" (UID: "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.977615 4862 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.977682 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-scripts" (OuterVolumeSpecName: "scripts") pod "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" (UID: "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.977689 4862 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.977728 4862 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:55 crc kubenswrapper[4862]: I1128 11:42:55.983225 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-kube-api-access-m87wm" (OuterVolumeSpecName: "kube-api-access-m87wm") pod "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" (UID: "a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c"). InnerVolumeSpecName "kube-api-access-m87wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:42:56 crc kubenswrapper[4862]: I1128 11:42:56.079024 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:56 crc kubenswrapper[4862]: I1128 11:42:56.079078 4862 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:56 crc kubenswrapper[4862]: I1128 11:42:56.079119 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m87wm\" (UniqueName: \"kubernetes.io/projected/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c-kube-api-access-m87wm\") on node \"crc\" DevicePath \"\"" Nov 28 11:42:56 crc kubenswrapper[4862]: I1128 11:42:56.497309 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hkgwr-config-jzcz8" event={"ID":"a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c","Type":"ContainerDied","Data":"5aa1d7841d9ffd14387be208cec6bd9f3daeef0b7c08dc4125cd4d06615a3b25"} Nov 28 11:42:56 crc kubenswrapper[4862]: I1128 11:42:56.497353 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hkgwr-config-jzcz8" Nov 28 11:42:56 crc kubenswrapper[4862]: I1128 11:42:56.497361 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5aa1d7841d9ffd14387be208cec6bd9f3daeef0b7c08dc4125cd4d06615a3b25" Nov 28 11:42:56 crc kubenswrapper[4862]: I1128 11:42:56.938085 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hkgwr-config-jzcz8"] Nov 28 11:42:56 crc kubenswrapper[4862]: I1128 11:42:56.948148 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-hkgwr-config-jzcz8"] Nov 28 11:42:58 crc kubenswrapper[4862]: I1128 11:42:58.847795 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" path="/var/lib/kubelet/pods/a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c/volumes" Nov 28 11:42:59 crc kubenswrapper[4862]: I1128 11:42:59.637962 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:59 crc kubenswrapper[4862]: I1128 11:42:59.646521 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"swift-storage-0\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " pod="openstack/swift-storage-0" Nov 28 11:42:59 crc kubenswrapper[4862]: I1128 11:42:59.944530 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 11:43:00 crc kubenswrapper[4862]: I1128 11:43:00.549017 4862 generic.go:334] "Generic (PLEG): container finished" podID="010a34d3-73a3-4f8e-9bc3-15bb24d1d41f" containerID="00badb80af0f696931c741191dda1d4b64ccf8e89732147e270eb8afbaa447c3" exitCode=0 Nov 28 11:43:00 crc kubenswrapper[4862]: I1128 11:43:00.549167 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qltsz" event={"ID":"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f","Type":"ContainerDied","Data":"00badb80af0f696931c741191dda1d4b64ccf8e89732147e270eb8afbaa447c3"} Nov 28 11:43:00 crc kubenswrapper[4862]: I1128 11:43:00.587735 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.169948 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.507823 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-8ndh9"] Nov 28 11:43:01 crc kubenswrapper[4862]: E1128 11:43:01.508152 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac70dca-3c3a-43fe-81ef-2677c476e9a4" containerName="swift-ring-rebalance" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.508182 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac70dca-3c3a-43fe-81ef-2677c476e9a4" containerName="swift-ring-rebalance" Nov 28 11:43:01 crc kubenswrapper[4862]: E1128 11:43:01.508204 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" containerName="ovn-config" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.508210 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" containerName="ovn-config" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.508358 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ac70dca-3c3a-43fe-81ef-2677c476e9a4" containerName="swift-ring-rebalance" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.508372 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6cfbfd1-9d5e-49b8-99ce-b6fd04ae077c" containerName="ovn-config" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.508860 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.519075 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-8ndh9"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.562663 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"7afc02dbfce6529a27ddb4ccc07fef4871e0b156b293bff22ddb94283f95f8c4"} Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.623544 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-nlmd8"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.624595 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.634616 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8d34-account-create-update-4jdj5"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.635476 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.637298 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.642505 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nlmd8"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.661691 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8d34-account-create-update-4jdj5"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.681288 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-operator-scripts\") pod \"cinder-db-create-8ndh9\" (UID: \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\") " pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.681422 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzv79\" (UniqueName: \"kubernetes.io/projected/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-kube-api-access-gzv79\") pod \"cinder-db-create-8ndh9\" (UID: \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\") " pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.713953 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-68e8-account-create-update-4zwm5"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.717664 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.720188 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.724988 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-68e8-account-create-update-4zwm5"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.772442 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-gmd9b"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.774133 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.777691 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.778043 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gn5r2" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.778104 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.778333 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.784961 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzv79\" (UniqueName: \"kubernetes.io/projected/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-kube-api-access-gzv79\") pod \"cinder-db-create-8ndh9\" (UID: \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\") " pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.785031 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-operator-scripts\") pod \"cinder-db-create-8ndh9\" (UID: \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\") " pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.785071 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adfa85c5-0924-4b39-bc3d-d2d884b19161-operator-scripts\") pod \"barbican-db-create-nlmd8\" (UID: \"adfa85c5-0924-4b39-bc3d-d2d884b19161\") " pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.785123 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5gg4\" (UniqueName: \"kubernetes.io/projected/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-kube-api-access-m5gg4\") pod \"barbican-8d34-account-create-update-4jdj5\" (UID: \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\") " pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.785153 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-operator-scripts\") pod \"barbican-8d34-account-create-update-4jdj5\" (UID: \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\") " pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.785178 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh8r5\" (UniqueName: \"kubernetes.io/projected/adfa85c5-0924-4b39-bc3d-d2d884b19161-kube-api-access-nh8r5\") pod \"barbican-db-create-nlmd8\" (UID: \"adfa85c5-0924-4b39-bc3d-d2d884b19161\") " pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.786329 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-operator-scripts\") pod \"cinder-db-create-8ndh9\" (UID: \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\") " pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.796292 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gmd9b"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.820266 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzv79\" (UniqueName: \"kubernetes.io/projected/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-kube-api-access-gzv79\") pod \"cinder-db-create-8ndh9\" (UID: \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\") " pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.825948 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.886817 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-config-data\") pod \"keystone-db-sync-gmd9b\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.886862 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbh9g\" (UniqueName: \"kubernetes.io/projected/d86c33d6-e42d-4544-af10-64f39f8201f7-kube-api-access-lbh9g\") pod \"cinder-68e8-account-create-update-4zwm5\" (UID: \"d86c33d6-e42d-4544-af10-64f39f8201f7\") " pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.886904 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adfa85c5-0924-4b39-bc3d-d2d884b19161-operator-scripts\") pod \"barbican-db-create-nlmd8\" (UID: \"adfa85c5-0924-4b39-bc3d-d2d884b19161\") " pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.886930 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgkt4\" (UniqueName: \"kubernetes.io/projected/5987550f-5082-4c09-a428-8ddc77c38598-kube-api-access-xgkt4\") pod \"keystone-db-sync-gmd9b\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.886961 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5gg4\" (UniqueName: \"kubernetes.io/projected/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-kube-api-access-m5gg4\") pod \"barbican-8d34-account-create-update-4jdj5\" (UID: \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\") " pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.886990 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-operator-scripts\") pod \"barbican-8d34-account-create-update-4jdj5\" (UID: \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\") " pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.887011 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-combined-ca-bundle\") pod \"keystone-db-sync-gmd9b\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.887034 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh8r5\" (UniqueName: \"kubernetes.io/projected/adfa85c5-0924-4b39-bc3d-d2d884b19161-kube-api-access-nh8r5\") pod \"barbican-db-create-nlmd8\" (UID: \"adfa85c5-0924-4b39-bc3d-d2d884b19161\") " pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.887063 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d86c33d6-e42d-4544-af10-64f39f8201f7-operator-scripts\") pod \"cinder-68e8-account-create-update-4zwm5\" (UID: \"d86c33d6-e42d-4544-af10-64f39f8201f7\") " pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.887846 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adfa85c5-0924-4b39-bc3d-d2d884b19161-operator-scripts\") pod \"barbican-db-create-nlmd8\" (UID: \"adfa85c5-0924-4b39-bc3d-d2d884b19161\") " pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.888007 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-operator-scripts\") pod \"barbican-8d34-account-create-update-4jdj5\" (UID: \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\") " pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.909106 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-7v76b"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.916787 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.922790 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7v76b"] Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.922916 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5gg4\" (UniqueName: \"kubernetes.io/projected/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-kube-api-access-m5gg4\") pod \"barbican-8d34-account-create-update-4jdj5\" (UID: \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\") " pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.936781 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh8r5\" (UniqueName: \"kubernetes.io/projected/adfa85c5-0924-4b39-bc3d-d2d884b19161-kube-api-access-nh8r5\") pod \"barbican-db-create-nlmd8\" (UID: \"adfa85c5-0924-4b39-bc3d-d2d884b19161\") " pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.956178 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.963765 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.988999 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-combined-ca-bundle\") pod \"keystone-db-sync-gmd9b\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.989063 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d86c33d6-e42d-4544-af10-64f39f8201f7-operator-scripts\") pod \"cinder-68e8-account-create-update-4zwm5\" (UID: \"d86c33d6-e42d-4544-af10-64f39f8201f7\") " pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.989104 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp7gs\" (UniqueName: \"kubernetes.io/projected/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-kube-api-access-hp7gs\") pod \"neutron-db-create-7v76b\" (UID: \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\") " pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.989144 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-operator-scripts\") pod \"neutron-db-create-7v76b\" (UID: \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\") " pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.989181 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-config-data\") pod \"keystone-db-sync-gmd9b\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.989207 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbh9g\" (UniqueName: \"kubernetes.io/projected/d86c33d6-e42d-4544-af10-64f39f8201f7-kube-api-access-lbh9g\") pod \"cinder-68e8-account-create-update-4zwm5\" (UID: \"d86c33d6-e42d-4544-af10-64f39f8201f7\") " pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.989234 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgkt4\" (UniqueName: \"kubernetes.io/projected/5987550f-5082-4c09-a428-8ddc77c38598-kube-api-access-xgkt4\") pod \"keystone-db-sync-gmd9b\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.990608 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d86c33d6-e42d-4544-af10-64f39f8201f7-operator-scripts\") pod \"cinder-68e8-account-create-update-4zwm5\" (UID: \"d86c33d6-e42d-4544-af10-64f39f8201f7\") " pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:01 crc kubenswrapper[4862]: I1128 11:43:01.999826 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-config-data\") pod \"keystone-db-sync-gmd9b\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.004740 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-combined-ca-bundle\") pod \"keystone-db-sync-gmd9b\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.016701 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbh9g\" (UniqueName: \"kubernetes.io/projected/d86c33d6-e42d-4544-af10-64f39f8201f7-kube-api-access-lbh9g\") pod \"cinder-68e8-account-create-update-4zwm5\" (UID: \"d86c33d6-e42d-4544-af10-64f39f8201f7\") " pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.029782 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgkt4\" (UniqueName: \"kubernetes.io/projected/5987550f-5082-4c09-a428-8ddc77c38598-kube-api-access-xgkt4\") pod \"keystone-db-sync-gmd9b\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.034840 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c4cb-account-create-update-fmhl6"] Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.036310 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.039803 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.040857 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.055507 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c4cb-account-create-update-fmhl6"] Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.092891 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qltsz" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.093925 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp7gs\" (UniqueName: \"kubernetes.io/projected/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-kube-api-access-hp7gs\") pod \"neutron-db-create-7v76b\" (UID: \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\") " pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.094109 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-operator-scripts\") pod \"neutron-db-create-7v76b\" (UID: \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\") " pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.095323 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-operator-scripts\") pod \"neutron-db-create-7v76b\" (UID: \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\") " pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.098437 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.133701 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp7gs\" (UniqueName: \"kubernetes.io/projected/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-kube-api-access-hp7gs\") pod \"neutron-db-create-7v76b\" (UID: \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\") " pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.195105 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-combined-ca-bundle\") pod \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.195255 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vfpp\" (UniqueName: \"kubernetes.io/projected/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-kube-api-access-6vfpp\") pod \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.195306 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-config-data\") pod \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.195332 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-db-sync-config-data\") pod \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\" (UID: \"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f\") " Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.195567 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aa74c33-9ed9-4d71-925e-e5140feebbff-operator-scripts\") pod \"neutron-c4cb-account-create-update-fmhl6\" (UID: \"3aa74c33-9ed9-4d71-925e-e5140feebbff\") " pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.195641 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zk6s\" (UniqueName: \"kubernetes.io/projected/3aa74c33-9ed9-4d71-925e-e5140feebbff-kube-api-access-2zk6s\") pod \"neutron-c4cb-account-create-update-fmhl6\" (UID: \"3aa74c33-9ed9-4d71-925e-e5140feebbff\") " pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.205585 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "010a34d3-73a3-4f8e-9bc3-15bb24d1d41f" (UID: "010a34d3-73a3-4f8e-9bc3-15bb24d1d41f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.207801 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-kube-api-access-6vfpp" (OuterVolumeSpecName: "kube-api-access-6vfpp") pod "010a34d3-73a3-4f8e-9bc3-15bb24d1d41f" (UID: "010a34d3-73a3-4f8e-9bc3-15bb24d1d41f"). InnerVolumeSpecName "kube-api-access-6vfpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.223258 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "010a34d3-73a3-4f8e-9bc3-15bb24d1d41f" (UID: "010a34d3-73a3-4f8e-9bc3-15bb24d1d41f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.250277 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-config-data" (OuterVolumeSpecName: "config-data") pod "010a34d3-73a3-4f8e-9bc3-15bb24d1d41f" (UID: "010a34d3-73a3-4f8e-9bc3-15bb24d1d41f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.297528 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aa74c33-9ed9-4d71-925e-e5140feebbff-operator-scripts\") pod \"neutron-c4cb-account-create-update-fmhl6\" (UID: \"3aa74c33-9ed9-4d71-925e-e5140feebbff\") " pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.297663 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zk6s\" (UniqueName: \"kubernetes.io/projected/3aa74c33-9ed9-4d71-925e-e5140feebbff-kube-api-access-2zk6s\") pod \"neutron-c4cb-account-create-update-fmhl6\" (UID: \"3aa74c33-9ed9-4d71-925e-e5140feebbff\") " pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.297770 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vfpp\" (UniqueName: \"kubernetes.io/projected/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-kube-api-access-6vfpp\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.297791 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.297804 4862 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.297816 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.298874 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aa74c33-9ed9-4d71-925e-e5140feebbff-operator-scripts\") pod \"neutron-c4cb-account-create-update-fmhl6\" (UID: \"3aa74c33-9ed9-4d71-925e-e5140feebbff\") " pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.314042 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.325002 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zk6s\" (UniqueName: \"kubernetes.io/projected/3aa74c33-9ed9-4d71-925e-e5140feebbff-kube-api-access-2zk6s\") pod \"neutron-c4cb-account-create-update-fmhl6\" (UID: \"3aa74c33-9ed9-4d71-925e-e5140feebbff\") " pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.383294 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.464845 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-68e8-account-create-update-4zwm5"] Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.482599 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-8ndh9"] Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.577215 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8d34-account-create-update-4jdj5"] Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.577533 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qltsz" event={"ID":"010a34d3-73a3-4f8e-9bc3-15bb24d1d41f","Type":"ContainerDied","Data":"efe828311882e481520b14baa1aa5981275605b518a440e5a64a12977ad00b70"} Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.577556 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efe828311882e481520b14baa1aa5981275605b518a440e5a64a12977ad00b70" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.577384 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qltsz" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.602842 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nlmd8"] Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.742632 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gmd9b"] Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.912915 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7v76b"] Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.916418 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7dd4845ddf-7ngnt"] Nov 28 11:43:02 crc kubenswrapper[4862]: E1128 11:43:02.916773 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="010a34d3-73a3-4f8e-9bc3-15bb24d1d41f" containerName="glance-db-sync" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.916791 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="010a34d3-73a3-4f8e-9bc3-15bb24d1d41f" containerName="glance-db-sync" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.916953 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="010a34d3-73a3-4f8e-9bc3-15bb24d1d41f" containerName="glance-db-sync" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.917755 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:02 crc kubenswrapper[4862]: I1128 11:43:02.950609 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dd4845ddf-7ngnt"] Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.010327 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-sb\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.010794 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-nb\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.010831 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7fbj\" (UniqueName: \"kubernetes.io/projected/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-kube-api-access-q7fbj\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.010895 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-dns-svc\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.010950 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-config\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.051576 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c4cb-account-create-update-fmhl6"] Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.112535 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-dns-svc\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.112595 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-config\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.112655 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-sb\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.112769 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-nb\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.112802 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7fbj\" (UniqueName: \"kubernetes.io/projected/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-kube-api-access-q7fbj\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.117913 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-config\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.129934 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-dns-svc\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.129990 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-sb\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.137997 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-nb\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.155917 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7fbj\" (UniqueName: \"kubernetes.io/projected/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-kube-api-access-q7fbj\") pod \"dnsmasq-dns-7dd4845ddf-7ngnt\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:03 crc kubenswrapper[4862]: I1128 11:43:03.271939 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:04 crc kubenswrapper[4862]: W1128 11:43:04.075702 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8abae65f_f5e4_4b16_a8f0_8892145fd6bf.slice/crio-63124a5247c4faff7d504b7495bc9b6a8a57c945fabca0767edada6cd62f4d44 WatchSource:0}: Error finding container 63124a5247c4faff7d504b7495bc9b6a8a57c945fabca0767edada6cd62f4d44: Status 404 returned error can't find the container with id 63124a5247c4faff7d504b7495bc9b6a8a57c945fabca0767edada6cd62f4d44 Nov 28 11:43:04 crc kubenswrapper[4862]: W1128 11:43:04.077477 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5987550f_5082_4c09_a428_8ddc77c38598.slice/crio-9a0a1250502b80d734727fc3a152a32d3675638d2d5692224257cc8ee64a8df3 WatchSource:0}: Error finding container 9a0a1250502b80d734727fc3a152a32d3675638d2d5692224257cc8ee64a8df3: Status 404 returned error can't find the container with id 9a0a1250502b80d734727fc3a152a32d3675638d2d5692224257cc8ee64a8df3 Nov 28 11:43:04 crc kubenswrapper[4862]: W1128 11:43:04.081205 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadfa85c5_0924_4b39_bc3d_d2d884b19161.slice/crio-041ed540535028248b3e388750ac88033a8b08bdb34029244d906d6cb796a47c WatchSource:0}: Error finding container 041ed540535028248b3e388750ac88033a8b08bdb34029244d906d6cb796a47c: Status 404 returned error can't find the container with id 041ed540535028248b3e388750ac88033a8b08bdb34029244d906d6cb796a47c Nov 28 11:43:04 crc kubenswrapper[4862]: W1128 11:43:04.085869 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd86c33d6_e42d_4544_af10_64f39f8201f7.slice/crio-fdf633ddb980a46179cea6f685081e4c6c44e4b8ada9fcc5d2f4b33d3b884d3e WatchSource:0}: Error finding container fdf633ddb980a46179cea6f685081e4c6c44e4b8ada9fcc5d2f4b33d3b884d3e: Status 404 returned error can't find the container with id fdf633ddb980a46179cea6f685081e4c6c44e4b8ada9fcc5d2f4b33d3b884d3e Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.135409 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.136214 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.136939 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.594274 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8ndh9" event={"ID":"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4","Type":"ContainerStarted","Data":"0f28ad331cb738df8055d1c295bfeb6493d33b47ddd7640edef5f7af8a83c564"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.594624 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8ndh9" event={"ID":"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4","Type":"ContainerStarted","Data":"af3eb274f883f80c5a0ad091ff785e8e253a409b16901dc5f17f1735ec5e4a80"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.597381 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.599335 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7v76b" event={"ID":"8abae65f-f5e4-4b16-a8f0-8892145fd6bf","Type":"ContainerStarted","Data":"fb97e0ded23986063a8aee20ab8624987a936bd554e2e3bf8bd543c73eed9c7a"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.599370 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7v76b" event={"ID":"8abae65f-f5e4-4b16-a8f0-8892145fd6bf","Type":"ContainerStarted","Data":"63124a5247c4faff7d504b7495bc9b6a8a57c945fabca0767edada6cd62f4d44"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.600319 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dd4845ddf-7ngnt"] Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.604149 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-68e8-account-create-update-4zwm5" event={"ID":"d86c33d6-e42d-4544-af10-64f39f8201f7","Type":"ContainerStarted","Data":"96915e9686bbfe96431db94d018ab010ba0fdc714bbd5170e37a594dd24d598b"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.604195 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-68e8-account-create-update-4zwm5" event={"ID":"d86c33d6-e42d-4544-af10-64f39f8201f7","Type":"ContainerStarted","Data":"fdf633ddb980a46179cea6f685081e4c6c44e4b8ada9fcc5d2f4b33d3b884d3e"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.607199 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nlmd8" event={"ID":"adfa85c5-0924-4b39-bc3d-d2d884b19161","Type":"ContainerStarted","Data":"ac31c3f95462a4c5062acc4cdb0ced218a95b344159616b29747888ef3fb38b8"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.607240 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nlmd8" event={"ID":"adfa85c5-0924-4b39-bc3d-d2d884b19161","Type":"ContainerStarted","Data":"041ed540535028248b3e388750ac88033a8b08bdb34029244d906d6cb796a47c"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.609709 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8d34-account-create-update-4jdj5" event={"ID":"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d","Type":"ContainerStarted","Data":"f55bf280b2e3933ff62cacb88d6bf607a6df0aff1ec46aa2f1e8a94ca25b0425"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.609748 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8d34-account-create-update-4jdj5" event={"ID":"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d","Type":"ContainerStarted","Data":"34066d819cd66a934d42a6577c8ca449e7fa512622e0d52d4d9d86510cf2a85e"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.612103 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c4cb-account-create-update-fmhl6" event={"ID":"3aa74c33-9ed9-4d71-925e-e5140feebbff","Type":"ContainerStarted","Data":"d98f2a5cfade1b9e22b1a5356a69fd726d97f61e64058fa2b2bf4d0f31c331b9"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.612149 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c4cb-account-create-update-fmhl6" event={"ID":"3aa74c33-9ed9-4d71-925e-e5140feebbff","Type":"ContainerStarted","Data":"16f470bd7b9564f5ceda46c2e5a7fe49a1de3e3dab7fa0d5db7ce9a68632d65c"} Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.614478 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gmd9b" event={"ID":"5987550f-5082-4c09-a428-8ddc77c38598","Type":"ContainerStarted","Data":"9a0a1250502b80d734727fc3a152a32d3675638d2d5692224257cc8ee64a8df3"} Nov 28 11:43:04 crc kubenswrapper[4862]: W1128 11:43:04.622950 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc657ab4a_f20b_46a1_aa1b_35aed553d6c0.slice/crio-d6de6a66a66e33666b4b81bcb0a6452ef09947628ef5a084c0f836a4f903e2d6 WatchSource:0}: Error finding container d6de6a66a66e33666b4b81bcb0a6452ef09947628ef5a084c0f836a4f903e2d6: Status 404 returned error can't find the container with id d6de6a66a66e33666b4b81bcb0a6452ef09947628ef5a084c0f836a4f903e2d6 Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.634828 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-8ndh9" podStartSLOduration=3.634811473 podStartE2EDuration="3.634811473s" podCreationTimestamp="2025-11-28 11:43:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:04.61852242 +0000 UTC m=+1262.151036351" watchObservedRunningTime="2025-11-28 11:43:04.634811473 +0000 UTC m=+1262.167325384" Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.642598 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-nlmd8" podStartSLOduration=3.6425802689999998 podStartE2EDuration="3.642580269s" podCreationTimestamp="2025-11-28 11:43:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:04.635220464 +0000 UTC m=+1262.167734385" watchObservedRunningTime="2025-11-28 11:43:04.642580269 +0000 UTC m=+1262.175094190" Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.662364 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-8d34-account-create-update-4jdj5" podStartSLOduration=3.6623435840000003 podStartE2EDuration="3.662343584s" podCreationTimestamp="2025-11-28 11:43:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:04.654148986 +0000 UTC m=+1262.186662907" watchObservedRunningTime="2025-11-28 11:43:04.662343584 +0000 UTC m=+1262.194857515" Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.671239 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c4cb-account-create-update-fmhl6" podStartSLOduration=2.6712279 podStartE2EDuration="2.6712279s" podCreationTimestamp="2025-11-28 11:43:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:04.669191075 +0000 UTC m=+1262.201704996" watchObservedRunningTime="2025-11-28 11:43:04.6712279 +0000 UTC m=+1262.203741821" Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.719715 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-7v76b" podStartSLOduration=3.719689926 podStartE2EDuration="3.719689926s" podCreationTimestamp="2025-11-28 11:43:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:04.692280609 +0000 UTC m=+1262.224794530" watchObservedRunningTime="2025-11-28 11:43:04.719689926 +0000 UTC m=+1262.252203847" Nov 28 11:43:04 crc kubenswrapper[4862]: I1128 11:43:04.722611 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-68e8-account-create-update-4zwm5" podStartSLOduration=3.7225989029999997 podStartE2EDuration="3.722598903s" podCreationTimestamp="2025-11-28 11:43:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:04.711122009 +0000 UTC m=+1262.243635930" watchObservedRunningTime="2025-11-28 11:43:04.722598903 +0000 UTC m=+1262.255112824" Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.630392 4862 generic.go:334] "Generic (PLEG): container finished" podID="adfa85c5-0924-4b39-bc3d-d2d884b19161" containerID="ac31c3f95462a4c5062acc4cdb0ced218a95b344159616b29747888ef3fb38b8" exitCode=0 Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.630816 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nlmd8" event={"ID":"adfa85c5-0924-4b39-bc3d-d2d884b19161","Type":"ContainerDied","Data":"ac31c3f95462a4c5062acc4cdb0ced218a95b344159616b29747888ef3fb38b8"} Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.636043 4862 generic.go:334] "Generic (PLEG): container finished" podID="f558e8a3-94d7-4bb8-a390-71d7ed87bb9d" containerID="f55bf280b2e3933ff62cacb88d6bf607a6df0aff1ec46aa2f1e8a94ca25b0425" exitCode=0 Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.636117 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8d34-account-create-update-4jdj5" event={"ID":"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d","Type":"ContainerDied","Data":"f55bf280b2e3933ff62cacb88d6bf607a6df0aff1ec46aa2f1e8a94ca25b0425"} Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.638033 4862 generic.go:334] "Generic (PLEG): container finished" podID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerID="937b64afb7bb9f4152b8a58bafebe35f318d99aa762feb8feb4e8780be7a9b02" exitCode=0 Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.638116 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" event={"ID":"c657ab4a-f20b-46a1-aa1b-35aed553d6c0","Type":"ContainerDied","Data":"937b64afb7bb9f4152b8a58bafebe35f318d99aa762feb8feb4e8780be7a9b02"} Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.638140 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" event={"ID":"c657ab4a-f20b-46a1-aa1b-35aed553d6c0","Type":"ContainerStarted","Data":"d6de6a66a66e33666b4b81bcb0a6452ef09947628ef5a084c0f836a4f903e2d6"} Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.640939 4862 generic.go:334] "Generic (PLEG): container finished" podID="3aa74c33-9ed9-4d71-925e-e5140feebbff" containerID="d98f2a5cfade1b9e22b1a5356a69fd726d97f61e64058fa2b2bf4d0f31c331b9" exitCode=0 Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.641207 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c4cb-account-create-update-fmhl6" event={"ID":"3aa74c33-9ed9-4d71-925e-e5140feebbff","Type":"ContainerDied","Data":"d98f2a5cfade1b9e22b1a5356a69fd726d97f61e64058fa2b2bf4d0f31c331b9"} Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.650036 4862 generic.go:334] "Generic (PLEG): container finished" podID="4a1ff8a7-99a3-4655-b491-0a10f01c2ee4" containerID="0f28ad331cb738df8055d1c295bfeb6493d33b47ddd7640edef5f7af8a83c564" exitCode=0 Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.650121 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8ndh9" event={"ID":"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4","Type":"ContainerDied","Data":"0f28ad331cb738df8055d1c295bfeb6493d33b47ddd7640edef5f7af8a83c564"} Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.660480 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3"} Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.660535 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc"} Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.663013 4862 generic.go:334] "Generic (PLEG): container finished" podID="8abae65f-f5e4-4b16-a8f0-8892145fd6bf" containerID="fb97e0ded23986063a8aee20ab8624987a936bd554e2e3bf8bd543c73eed9c7a" exitCode=0 Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.663109 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7v76b" event={"ID":"8abae65f-f5e4-4b16-a8f0-8892145fd6bf","Type":"ContainerDied","Data":"fb97e0ded23986063a8aee20ab8624987a936bd554e2e3bf8bd543c73eed9c7a"} Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.665134 4862 generic.go:334] "Generic (PLEG): container finished" podID="d86c33d6-e42d-4544-af10-64f39f8201f7" containerID="96915e9686bbfe96431db94d018ab010ba0fdc714bbd5170e37a594dd24d598b" exitCode=0 Nov 28 11:43:05 crc kubenswrapper[4862]: I1128 11:43:05.665175 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-68e8-account-create-update-4zwm5" event={"ID":"d86c33d6-e42d-4544-af10-64f39f8201f7","Type":"ContainerDied","Data":"96915e9686bbfe96431db94d018ab010ba0fdc714bbd5170e37a594dd24d598b"} Nov 28 11:43:06 crc kubenswrapper[4862]: I1128 11:43:06.677136 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a"} Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.091025 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.195368 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-operator-scripts\") pod \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\" (UID: \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.195440 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp7gs\" (UniqueName: \"kubernetes.io/projected/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-kube-api-access-hp7gs\") pod \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\" (UID: \"8abae65f-f5e4-4b16-a8f0-8892145fd6bf\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.197350 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8abae65f-f5e4-4b16-a8f0-8892145fd6bf" (UID: "8abae65f-f5e4-4b16-a8f0-8892145fd6bf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.210472 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-kube-api-access-hp7gs" (OuterVolumeSpecName: "kube-api-access-hp7gs") pod "8abae65f-f5e4-4b16-a8f0-8892145fd6bf" (UID: "8abae65f-f5e4-4b16-a8f0-8892145fd6bf"). InnerVolumeSpecName "kube-api-access-hp7gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.297774 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp7gs\" (UniqueName: \"kubernetes.io/projected/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-kube-api-access-hp7gs\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.297803 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8abae65f-f5e4-4b16-a8f0-8892145fd6bf-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.344845 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.351162 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.355325 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.369393 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.374206 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.399112 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbh9g\" (UniqueName: \"kubernetes.io/projected/d86c33d6-e42d-4544-af10-64f39f8201f7-kube-api-access-lbh9g\") pod \"d86c33d6-e42d-4544-af10-64f39f8201f7\" (UID: \"d86c33d6-e42d-4544-af10-64f39f8201f7\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.399586 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d86c33d6-e42d-4544-af10-64f39f8201f7-operator-scripts\") pod \"d86c33d6-e42d-4544-af10-64f39f8201f7\" (UID: \"d86c33d6-e42d-4544-af10-64f39f8201f7\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.400477 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d86c33d6-e42d-4544-af10-64f39f8201f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d86c33d6-e42d-4544-af10-64f39f8201f7" (UID: "d86c33d6-e42d-4544-af10-64f39f8201f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.403317 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d86c33d6-e42d-4544-af10-64f39f8201f7-kube-api-access-lbh9g" (OuterVolumeSpecName: "kube-api-access-lbh9g") pod "d86c33d6-e42d-4544-af10-64f39f8201f7" (UID: "d86c33d6-e42d-4544-af10-64f39f8201f7"). InnerVolumeSpecName "kube-api-access-lbh9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501175 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aa74c33-9ed9-4d71-925e-e5140feebbff-operator-scripts\") pod \"3aa74c33-9ed9-4d71-925e-e5140feebbff\" (UID: \"3aa74c33-9ed9-4d71-925e-e5140feebbff\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501266 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5gg4\" (UniqueName: \"kubernetes.io/projected/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-kube-api-access-m5gg4\") pod \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\" (UID: \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501309 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzv79\" (UniqueName: \"kubernetes.io/projected/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-kube-api-access-gzv79\") pod \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\" (UID: \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501390 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh8r5\" (UniqueName: \"kubernetes.io/projected/adfa85c5-0924-4b39-bc3d-d2d884b19161-kube-api-access-nh8r5\") pod \"adfa85c5-0924-4b39-bc3d-d2d884b19161\" (UID: \"adfa85c5-0924-4b39-bc3d-d2d884b19161\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501407 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adfa85c5-0924-4b39-bc3d-d2d884b19161-operator-scripts\") pod \"adfa85c5-0924-4b39-bc3d-d2d884b19161\" (UID: \"adfa85c5-0924-4b39-bc3d-d2d884b19161\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501431 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-operator-scripts\") pod \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\" (UID: \"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501486 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zk6s\" (UniqueName: \"kubernetes.io/projected/3aa74c33-9ed9-4d71-925e-e5140feebbff-kube-api-access-2zk6s\") pod \"3aa74c33-9ed9-4d71-925e-e5140feebbff\" (UID: \"3aa74c33-9ed9-4d71-925e-e5140feebbff\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501522 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-operator-scripts\") pod \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\" (UID: \"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4\") " Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501894 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d86c33d6-e42d-4544-af10-64f39f8201f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.501914 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbh9g\" (UniqueName: \"kubernetes.io/projected/d86c33d6-e42d-4544-af10-64f39f8201f7-kube-api-access-lbh9g\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.502331 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4a1ff8a7-99a3-4655-b491-0a10f01c2ee4" (UID: "4a1ff8a7-99a3-4655-b491-0a10f01c2ee4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.502329 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f558e8a3-94d7-4bb8-a390-71d7ed87bb9d" (UID: "f558e8a3-94d7-4bb8-a390-71d7ed87bb9d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.502399 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa74c33-9ed9-4d71-925e-e5140feebbff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3aa74c33-9ed9-4d71-925e-e5140feebbff" (UID: "3aa74c33-9ed9-4d71-925e-e5140feebbff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.502594 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adfa85c5-0924-4b39-bc3d-d2d884b19161-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "adfa85c5-0924-4b39-bc3d-d2d884b19161" (UID: "adfa85c5-0924-4b39-bc3d-d2d884b19161"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.505193 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adfa85c5-0924-4b39-bc3d-d2d884b19161-kube-api-access-nh8r5" (OuterVolumeSpecName: "kube-api-access-nh8r5") pod "adfa85c5-0924-4b39-bc3d-d2d884b19161" (UID: "adfa85c5-0924-4b39-bc3d-d2d884b19161"). InnerVolumeSpecName "kube-api-access-nh8r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.505312 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-kube-api-access-m5gg4" (OuterVolumeSpecName: "kube-api-access-m5gg4") pod "f558e8a3-94d7-4bb8-a390-71d7ed87bb9d" (UID: "f558e8a3-94d7-4bb8-a390-71d7ed87bb9d"). InnerVolumeSpecName "kube-api-access-m5gg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.505357 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-kube-api-access-gzv79" (OuterVolumeSpecName: "kube-api-access-gzv79") pod "4a1ff8a7-99a3-4655-b491-0a10f01c2ee4" (UID: "4a1ff8a7-99a3-4655-b491-0a10f01c2ee4"). InnerVolumeSpecName "kube-api-access-gzv79". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.507661 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa74c33-9ed9-4d71-925e-e5140feebbff-kube-api-access-2zk6s" (OuterVolumeSpecName: "kube-api-access-2zk6s") pod "3aa74c33-9ed9-4d71-925e-e5140feebbff" (UID: "3aa74c33-9ed9-4d71-925e-e5140feebbff"). InnerVolumeSpecName "kube-api-access-2zk6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.604151 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzv79\" (UniqueName: \"kubernetes.io/projected/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-kube-api-access-gzv79\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.604205 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh8r5\" (UniqueName: \"kubernetes.io/projected/adfa85c5-0924-4b39-bc3d-d2d884b19161-kube-api-access-nh8r5\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.604230 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adfa85c5-0924-4b39-bc3d-d2d884b19161-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.604247 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.604265 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zk6s\" (UniqueName: \"kubernetes.io/projected/3aa74c33-9ed9-4d71-925e-e5140feebbff-kube-api-access-2zk6s\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.604282 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.604301 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aa74c33-9ed9-4d71-925e-e5140feebbff-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.604319 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5gg4\" (UniqueName: \"kubernetes.io/projected/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d-kube-api-access-m5gg4\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.696854 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7v76b" event={"ID":"8abae65f-f5e4-4b16-a8f0-8892145fd6bf","Type":"ContainerDied","Data":"63124a5247c4faff7d504b7495bc9b6a8a57c945fabca0767edada6cd62f4d44"} Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.696903 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63124a5247c4faff7d504b7495bc9b6a8a57c945fabca0767edada6cd62f4d44" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.696970 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7v76b" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.698783 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-68e8-account-create-update-4zwm5" event={"ID":"d86c33d6-e42d-4544-af10-64f39f8201f7","Type":"ContainerDied","Data":"fdf633ddb980a46179cea6f685081e4c6c44e4b8ada9fcc5d2f4b33d3b884d3e"} Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.698825 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdf633ddb980a46179cea6f685081e4c6c44e4b8ada9fcc5d2f4b33d3b884d3e" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.699008 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-68e8-account-create-update-4zwm5" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.700198 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nlmd8" event={"ID":"adfa85c5-0924-4b39-bc3d-d2d884b19161","Type":"ContainerDied","Data":"041ed540535028248b3e388750ac88033a8b08bdb34029244d906d6cb796a47c"} Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.700238 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="041ed540535028248b3e388750ac88033a8b08bdb34029244d906d6cb796a47c" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.700216 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nlmd8" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.701513 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8d34-account-create-update-4jdj5" event={"ID":"f558e8a3-94d7-4bb8-a390-71d7ed87bb9d","Type":"ContainerDied","Data":"34066d819cd66a934d42a6577c8ca449e7fa512622e0d52d4d9d86510cf2a85e"} Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.701538 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34066d819cd66a934d42a6577c8ca449e7fa512622e0d52d4d9d86510cf2a85e" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.701590 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8d34-account-create-update-4jdj5" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.713154 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c4cb-account-create-update-fmhl6" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.713266 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c4cb-account-create-update-fmhl6" event={"ID":"3aa74c33-9ed9-4d71-925e-e5140feebbff","Type":"ContainerDied","Data":"16f470bd7b9564f5ceda46c2e5a7fe49a1de3e3dab7fa0d5db7ce9a68632d65c"} Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.713295 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16f470bd7b9564f5ceda46c2e5a7fe49a1de3e3dab7fa0d5db7ce9a68632d65c" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.716629 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8ndh9" event={"ID":"4a1ff8a7-99a3-4655-b491-0a10f01c2ee4","Type":"ContainerDied","Data":"af3eb274f883f80c5a0ad091ff785e8e253a409b16901dc5f17f1735ec5e4a80"} Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.716671 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af3eb274f883f80c5a0ad091ff785e8e253a409b16901dc5f17f1735ec5e4a80" Nov 28 11:43:07 crc kubenswrapper[4862]: I1128 11:43:07.716797 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8ndh9" Nov 28 11:43:09 crc kubenswrapper[4862]: I1128 11:43:09.740331 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" event={"ID":"c657ab4a-f20b-46a1-aa1b-35aed553d6c0","Type":"ContainerStarted","Data":"d16ba579580174081c4341491a21bdbf0d1f32acd583f99f26471ced2f0fe306"} Nov 28 11:43:10 crc kubenswrapper[4862]: I1128 11:43:10.753261 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:10 crc kubenswrapper[4862]: I1128 11:43:10.794255 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" podStartSLOduration=8.794232567 podStartE2EDuration="8.794232567s" podCreationTimestamp="2025-11-28 11:43:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:10.792724987 +0000 UTC m=+1268.325238908" watchObservedRunningTime="2025-11-28 11:43:10.794232567 +0000 UTC m=+1268.326746508" Nov 28 11:43:11 crc kubenswrapper[4862]: I1128 11:43:11.443395 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:43:14 crc kubenswrapper[4862]: I1128 11:43:14.785042 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gmd9b" event={"ID":"5987550f-5082-4c09-a428-8ddc77c38598","Type":"ContainerStarted","Data":"8796e3b178d354b282dda262ad4d4df0cd3fe71afbe530ecd93777cdc96c0188"} Nov 28 11:43:14 crc kubenswrapper[4862]: I1128 11:43:14.787474 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235"} Nov 28 11:43:14 crc kubenswrapper[4862]: I1128 11:43:14.815734 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-gmd9b" podStartSLOduration=4.120942199 podStartE2EDuration="13.815717898s" podCreationTimestamp="2025-11-28 11:43:01 +0000 UTC" firstStartedPulling="2025-11-28 11:43:04.080828284 +0000 UTC m=+1261.613342205" lastFinishedPulling="2025-11-28 11:43:13.775603953 +0000 UTC m=+1271.308117904" observedRunningTime="2025-11-28 11:43:14.811153197 +0000 UTC m=+1272.343667118" watchObservedRunningTime="2025-11-28 11:43:14.815717898 +0000 UTC m=+1272.348231819" Nov 28 11:43:15 crc kubenswrapper[4862]: I1128 11:43:15.805691 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726"} Nov 28 11:43:15 crc kubenswrapper[4862]: I1128 11:43:15.805785 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a"} Nov 28 11:43:15 crc kubenswrapper[4862]: I1128 11:43:15.805806 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61"} Nov 28 11:43:17 crc kubenswrapper[4862]: I1128 11:43:17.830315 4862 generic.go:334] "Generic (PLEG): container finished" podID="5987550f-5082-4c09-a428-8ddc77c38598" containerID="8796e3b178d354b282dda262ad4d4df0cd3fe71afbe530ecd93777cdc96c0188" exitCode=0 Nov 28 11:43:17 crc kubenswrapper[4862]: I1128 11:43:17.830363 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gmd9b" event={"ID":"5987550f-5082-4c09-a428-8ddc77c38598","Type":"ContainerDied","Data":"8796e3b178d354b282dda262ad4d4df0cd3fe71afbe530ecd93777cdc96c0188"} Nov 28 11:43:17 crc kubenswrapper[4862]: I1128 11:43:17.840415 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d"} Nov 28 11:43:17 crc kubenswrapper[4862]: I1128 11:43:17.840465 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97"} Nov 28 11:43:17 crc kubenswrapper[4862]: I1128 11:43:17.840476 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c"} Nov 28 11:43:17 crc kubenswrapper[4862]: I1128 11:43:17.840487 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02"} Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.275335 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.349683 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c8cb8df65-lvvcw"] Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.349953 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" podUID="966949c3-c79d-4811-8c75-c4e913cd6a19" containerName="dnsmasq-dns" containerID="cri-o://97c0b76d1ae2e41b6a809bf0f557b15892c29ab5beabbc51aa8dc791239b70cb" gracePeriod=10 Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.863661 4862 generic.go:334] "Generic (PLEG): container finished" podID="966949c3-c79d-4811-8c75-c4e913cd6a19" containerID="97c0b76d1ae2e41b6a809bf0f557b15892c29ab5beabbc51aa8dc791239b70cb" exitCode=0 Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.863735 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" event={"ID":"966949c3-c79d-4811-8c75-c4e913cd6a19","Type":"ContainerDied","Data":"97c0b76d1ae2e41b6a809bf0f557b15892c29ab5beabbc51aa8dc791239b70cb"} Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.864057 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" event={"ID":"966949c3-c79d-4811-8c75-c4e913cd6a19","Type":"ContainerDied","Data":"2b6ea71ea113385a11cc58c86cddd888e3de762caa0ed7a8592c00822a2db0a3"} Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.864067 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b6ea71ea113385a11cc58c86cddd888e3de762caa0ed7a8592c00822a2db0a3" Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.874186 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255"} Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.874221 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6"} Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.874232 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerStarted","Data":"0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3"} Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.878654 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:43:18 crc kubenswrapper[4862]: I1128 11:43:18.914707 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.805685298 podStartE2EDuration="52.914668776s" podCreationTimestamp="2025-11-28 11:42:26 +0000 UTC" firstStartedPulling="2025-11-28 11:43:00.597016238 +0000 UTC m=+1258.129530159" lastFinishedPulling="2025-11-28 11:43:16.705999716 +0000 UTC m=+1274.238513637" observedRunningTime="2025-11-28 11:43:18.913586048 +0000 UTC m=+1276.446099969" watchObservedRunningTime="2025-11-28 11:43:18.914668776 +0000 UTC m=+1276.447182698" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.020795 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-sb\") pod \"966949c3-c79d-4811-8c75-c4e913cd6a19\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.020855 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd92k\" (UniqueName: \"kubernetes.io/projected/966949c3-c79d-4811-8c75-c4e913cd6a19-kube-api-access-bd92k\") pod \"966949c3-c79d-4811-8c75-c4e913cd6a19\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.021170 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-dns-svc\") pod \"966949c3-c79d-4811-8c75-c4e913cd6a19\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.021234 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-config\") pod \"966949c3-c79d-4811-8c75-c4e913cd6a19\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.021276 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-nb\") pod \"966949c3-c79d-4811-8c75-c4e913cd6a19\" (UID: \"966949c3-c79d-4811-8c75-c4e913cd6a19\") " Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.039468 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/966949c3-c79d-4811-8c75-c4e913cd6a19-kube-api-access-bd92k" (OuterVolumeSpecName: "kube-api-access-bd92k") pod "966949c3-c79d-4811-8c75-c4e913cd6a19" (UID: "966949c3-c79d-4811-8c75-c4e913cd6a19"). InnerVolumeSpecName "kube-api-access-bd92k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.098907 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "966949c3-c79d-4811-8c75-c4e913cd6a19" (UID: "966949c3-c79d-4811-8c75-c4e913cd6a19"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.101812 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "966949c3-c79d-4811-8c75-c4e913cd6a19" (UID: "966949c3-c79d-4811-8c75-c4e913cd6a19"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.108766 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-config" (OuterVolumeSpecName: "config") pod "966949c3-c79d-4811-8c75-c4e913cd6a19" (UID: "966949c3-c79d-4811-8c75-c4e913cd6a19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.117145 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "966949c3-c79d-4811-8c75-c4e913cd6a19" (UID: "966949c3-c79d-4811-8c75-c4e913cd6a19"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.124755 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.124780 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.124791 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.124801 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/966949c3-c79d-4811-8c75-c4e913cd6a19-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.124811 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd92k\" (UniqueName: \"kubernetes.io/projected/966949c3-c79d-4811-8c75-c4e913cd6a19-kube-api-access-bd92k\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.174427 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b6bbf7467-wd82l"] Nov 28 11:43:19 crc kubenswrapper[4862]: E1128 11:43:19.174953 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa74c33-9ed9-4d71-925e-e5140feebbff" containerName="mariadb-account-create-update" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.174970 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa74c33-9ed9-4d71-925e-e5140feebbff" containerName="mariadb-account-create-update" Nov 28 11:43:19 crc kubenswrapper[4862]: E1128 11:43:19.174983 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="966949c3-c79d-4811-8c75-c4e913cd6a19" containerName="init" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.174990 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="966949c3-c79d-4811-8c75-c4e913cd6a19" containerName="init" Nov 28 11:43:19 crc kubenswrapper[4862]: E1128 11:43:19.175003 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a1ff8a7-99a3-4655-b491-0a10f01c2ee4" containerName="mariadb-database-create" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175009 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a1ff8a7-99a3-4655-b491-0a10f01c2ee4" containerName="mariadb-database-create" Nov 28 11:43:19 crc kubenswrapper[4862]: E1128 11:43:19.175019 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="966949c3-c79d-4811-8c75-c4e913cd6a19" containerName="dnsmasq-dns" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175024 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="966949c3-c79d-4811-8c75-c4e913cd6a19" containerName="dnsmasq-dns" Nov 28 11:43:19 crc kubenswrapper[4862]: E1128 11:43:19.175040 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f558e8a3-94d7-4bb8-a390-71d7ed87bb9d" containerName="mariadb-account-create-update" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175046 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f558e8a3-94d7-4bb8-a390-71d7ed87bb9d" containerName="mariadb-account-create-update" Nov 28 11:43:19 crc kubenswrapper[4862]: E1128 11:43:19.175054 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8abae65f-f5e4-4b16-a8f0-8892145fd6bf" containerName="mariadb-database-create" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175061 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="8abae65f-f5e4-4b16-a8f0-8892145fd6bf" containerName="mariadb-database-create" Nov 28 11:43:19 crc kubenswrapper[4862]: E1128 11:43:19.175075 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adfa85c5-0924-4b39-bc3d-d2d884b19161" containerName="mariadb-database-create" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175081 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="adfa85c5-0924-4b39-bc3d-d2d884b19161" containerName="mariadb-database-create" Nov 28 11:43:19 crc kubenswrapper[4862]: E1128 11:43:19.175115 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d86c33d6-e42d-4544-af10-64f39f8201f7" containerName="mariadb-account-create-update" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175121 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="d86c33d6-e42d-4544-af10-64f39f8201f7" containerName="mariadb-account-create-update" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175264 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="adfa85c5-0924-4b39-bc3d-d2d884b19161" containerName="mariadb-database-create" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175283 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a1ff8a7-99a3-4655-b491-0a10f01c2ee4" containerName="mariadb-database-create" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175299 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="966949c3-c79d-4811-8c75-c4e913cd6a19" containerName="dnsmasq-dns" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175310 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa74c33-9ed9-4d71-925e-e5140feebbff" containerName="mariadb-account-create-update" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175320 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f558e8a3-94d7-4bb8-a390-71d7ed87bb9d" containerName="mariadb-account-create-update" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175327 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="8abae65f-f5e4-4b16-a8f0-8892145fd6bf" containerName="mariadb-database-create" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.175359 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="d86c33d6-e42d-4544-af10-64f39f8201f7" containerName="mariadb-account-create-update" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.176185 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.178197 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.189949 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6bbf7467-wd82l"] Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.190459 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.228149 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-config-data\") pod \"5987550f-5082-4c09-a428-8ddc77c38598\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.228286 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-combined-ca-bundle\") pod \"5987550f-5082-4c09-a428-8ddc77c38598\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.228399 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgkt4\" (UniqueName: \"kubernetes.io/projected/5987550f-5082-4c09-a428-8ddc77c38598-kube-api-access-xgkt4\") pod \"5987550f-5082-4c09-a428-8ddc77c38598\" (UID: \"5987550f-5082-4c09-a428-8ddc77c38598\") " Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.228888 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq8s9\" (UniqueName: \"kubernetes.io/projected/6cd448ac-1ecc-4de0-841e-08a1df0f1263-kube-api-access-bq8s9\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.228933 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-swift-storage-0\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.228953 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-sb\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.228982 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-nb\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.229020 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-svc\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.229044 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-config\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.234619 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5987550f-5082-4c09-a428-8ddc77c38598-kube-api-access-xgkt4" (OuterVolumeSpecName: "kube-api-access-xgkt4") pod "5987550f-5082-4c09-a428-8ddc77c38598" (UID: "5987550f-5082-4c09-a428-8ddc77c38598"). InnerVolumeSpecName "kube-api-access-xgkt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.259311 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5987550f-5082-4c09-a428-8ddc77c38598" (UID: "5987550f-5082-4c09-a428-8ddc77c38598"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.273812 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-config-data" (OuterVolumeSpecName: "config-data") pod "5987550f-5082-4c09-a428-8ddc77c38598" (UID: "5987550f-5082-4c09-a428-8ddc77c38598"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.330548 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-svc\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.330628 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-config\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.330816 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq8s9\" (UniqueName: \"kubernetes.io/projected/6cd448ac-1ecc-4de0-841e-08a1df0f1263-kube-api-access-bq8s9\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.330869 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-swift-storage-0\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.330902 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-sb\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.330954 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-nb\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.331594 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgkt4\" (UniqueName: \"kubernetes.io/projected/5987550f-5082-4c09-a428-8ddc77c38598-kube-api-access-xgkt4\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.331642 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.331666 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5987550f-5082-4c09-a428-8ddc77c38598-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.332327 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-config\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.332345 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-swift-storage-0\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.332731 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-nb\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.332899 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-sb\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.333449 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-svc\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.348773 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq8s9\" (UniqueName: \"kubernetes.io/projected/6cd448ac-1ecc-4de0-841e-08a1df0f1263-kube-api-access-bq8s9\") pod \"dnsmasq-dns-b6bbf7467-wd82l\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.528961 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.897340 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gmd9b" event={"ID":"5987550f-5082-4c09-a428-8ddc77c38598","Type":"ContainerDied","Data":"9a0a1250502b80d734727fc3a152a32d3675638d2d5692224257cc8ee64a8df3"} Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.897658 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a0a1250502b80d734727fc3a152a32d3675638d2d5692224257cc8ee64a8df3" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.897446 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cb8df65-lvvcw" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.897425 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gmd9b" Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.930052 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c8cb8df65-lvvcw"] Nov 28 11:43:19 crc kubenswrapper[4862]: I1128 11:43:19.938896 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c8cb8df65-lvvcw"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.071044 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6bbf7467-wd82l"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.099802 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6bbf7467-wd82l"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.156240 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-654fz"] Nov 28 11:43:20 crc kubenswrapper[4862]: E1128 11:43:20.157029 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5987550f-5082-4c09-a428-8ddc77c38598" containerName="keystone-db-sync" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.157051 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="5987550f-5082-4c09-a428-8ddc77c38598" containerName="keystone-db-sync" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.157321 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="5987550f-5082-4c09-a428-8ddc77c38598" containerName="keystone-db-sync" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.158077 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.163788 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.163939 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gn5r2" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.164762 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.169582 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.169866 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.177188 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76bb7864cf-2r6pg"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.178751 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.188706 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-654fz"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.200066 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76bb7864cf-2r6pg"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.256129 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl57w\" (UniqueName: \"kubernetes.io/projected/9cd16bba-3339-4b8e-b495-2a8a47c6a835-kube-api-access-kl57w\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.256207 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-combined-ca-bundle\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.256258 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-credential-keys\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.256276 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-nb\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.256322 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-scripts\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.256350 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-sb\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.256405 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-fernet-keys\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.256421 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-config\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.256454 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-config-data\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.257013 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-swift-storage-0\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.257043 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2swg\" (UniqueName: \"kubernetes.io/projected/de671dfb-c489-4f8c-8e77-da684cca1f98-kube-api-access-l2swg\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.257079 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-svc\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.358762 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl57w\" (UniqueName: \"kubernetes.io/projected/9cd16bba-3339-4b8e-b495-2a8a47c6a835-kube-api-access-kl57w\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.358825 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-combined-ca-bundle\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.358858 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-credential-keys\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.358877 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-nb\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.358909 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-scripts\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.358931 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-sb\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.358961 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-fernet-keys\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.358976 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-config\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.359000 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-config-data\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.359028 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-swift-storage-0\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.359044 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2swg\" (UniqueName: \"kubernetes.io/projected/de671dfb-c489-4f8c-8e77-da684cca1f98-kube-api-access-l2swg\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.359066 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-svc\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.359914 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-svc\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.364933 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-nb\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.369002 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-fernet-keys\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.369647 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-sb\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.369955 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-config\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.370234 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-swift-storage-0\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.370549 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-credential-keys\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.370741 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-combined-ca-bundle\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.371674 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-scripts\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.373237 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-config-data\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.388149 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.390015 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.396481 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.396666 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.406301 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2swg\" (UniqueName: \"kubernetes.io/projected/de671dfb-c489-4f8c-8e77-da684cca1f98-kube-api-access-l2swg\") pod \"dnsmasq-dns-76bb7864cf-2r6pg\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.414706 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl57w\" (UniqueName: \"kubernetes.io/projected/9cd16bba-3339-4b8e-b495-2a8a47c6a835-kube-api-access-kl57w\") pod \"keystone-bootstrap-654fz\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.423047 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.441764 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-5jcj5"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.442883 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.455242 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.455750 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.456456 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ggp6j" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.460973 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-scripts\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.461026 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pskx\" (UniqueName: \"kubernetes.io/projected/5c005a75-9925-4968-bb0f-40c16c805d35-kube-api-access-2pskx\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.461046 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-config-data\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.461074 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-log-httpd\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.461118 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-run-httpd\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.461162 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.461221 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.474158 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-47sh5"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.491935 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.493437 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5jcj5"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.508169 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.508256 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zbk6l" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.508496 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.536655 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-47sh5"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.562759 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.562827 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlgwn\" (UniqueName: \"kubernetes.io/projected/82ea93c1-317d-4d0f-92a6-93872ef50d12-kube-api-access-dlgwn\") pod \"neutron-db-sync-47sh5\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.562873 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b376a4e9-5456-4e8f-9c91-b222f38c30db-etc-machine-id\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.562913 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-config-data\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.562932 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-combined-ca-bundle\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.562957 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.562973 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-db-sync-config-data\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.562994 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xdmc\" (UniqueName: \"kubernetes.io/projected/b376a4e9-5456-4e8f-9c91-b222f38c30db-kube-api-access-5xdmc\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.563013 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-scripts\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.563029 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-config\") pod \"neutron-db-sync-47sh5\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.563050 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-combined-ca-bundle\") pod \"neutron-db-sync-47sh5\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.563073 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pskx\" (UniqueName: \"kubernetes.io/projected/5c005a75-9925-4968-bb0f-40c16c805d35-kube-api-access-2pskx\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.563102 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-config-data\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.563125 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-log-httpd\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.563156 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-run-httpd\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.563170 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-scripts\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.564292 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-log-httpd\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.564309 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-run-httpd\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.566862 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dnh4p"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.568103 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.571780 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-config-data\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.573689 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-nvwvc" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.574334 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.581359 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.582438 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-scripts\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.583043 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.594593 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dnh4p"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.600038 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pskx\" (UniqueName: \"kubernetes.io/projected/5c005a75-9925-4968-bb0f-40c16c805d35-kube-api-access-2pskx\") pod \"ceilometer-0\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.602464 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.617816 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.637362 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76bb7864cf-2r6pg"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.654402 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f5d458b55-bfmjn"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.656710 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.666778 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-scripts\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.666843 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-combined-ca-bundle\") pod \"barbican-db-sync-dnh4p\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.666893 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlgwn\" (UniqueName: \"kubernetes.io/projected/82ea93c1-317d-4d0f-92a6-93872ef50d12-kube-api-access-dlgwn\") pod \"neutron-db-sync-47sh5\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.666944 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b376a4e9-5456-4e8f-9c91-b222f38c30db-etc-machine-id\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.666968 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-config-data\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.666987 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-combined-ca-bundle\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.667012 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-db-sync-config-data\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.667034 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6pz5\" (UniqueName: \"kubernetes.io/projected/4be2cd5e-1345-4212-9563-449a665090f1-kube-api-access-g6pz5\") pod \"barbican-db-sync-dnh4p\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.667058 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xdmc\" (UniqueName: \"kubernetes.io/projected/b376a4e9-5456-4e8f-9c91-b222f38c30db-kube-api-access-5xdmc\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.667081 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-db-sync-config-data\") pod \"barbican-db-sync-dnh4p\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.670717 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-config\") pod \"neutron-db-sync-47sh5\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.670796 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-combined-ca-bundle\") pod \"neutron-db-sync-47sh5\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.674073 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b376a4e9-5456-4e8f-9c91-b222f38c30db-etc-machine-id\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.684339 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-db-sync-config-data\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.685755 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-config\") pod \"neutron-db-sync-47sh5\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.686395 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-combined-ca-bundle\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.696333 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xdmc\" (UniqueName: \"kubernetes.io/projected/b376a4e9-5456-4e8f-9c91-b222f38c30db-kube-api-access-5xdmc\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.698077 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-combined-ca-bundle\") pod \"neutron-db-sync-47sh5\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.698533 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-config-data\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.699423 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-scripts\") pod \"cinder-db-sync-5jcj5\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.702566 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlgwn\" (UniqueName: \"kubernetes.io/projected/82ea93c1-317d-4d0f-92a6-93872ef50d12-kube-api-access-dlgwn\") pod \"neutron-db-sync-47sh5\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.736958 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.741868 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-5lsz4"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.745824 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.749035 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.749316 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.749423 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-p6ckl" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.751388 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f5d458b55-bfmjn"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.758590 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5lsz4"] Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.801535 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.801588 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6pz5\" (UniqueName: \"kubernetes.io/projected/4be2cd5e-1345-4212-9563-449a665090f1-kube-api-access-g6pz5\") pod \"barbican-db-sync-dnh4p\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.801617 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-db-sync-config-data\") pod \"barbican-db-sync-dnh4p\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.801639 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-svc\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.801732 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.801750 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-combined-ca-bundle\") pod \"barbican-db-sync-dnh4p\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.801778 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr56q\" (UniqueName: \"kubernetes.io/projected/0bca720b-962c-48e5-9861-406c78a5b202-kube-api-access-rr56q\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.801816 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-config\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.801839 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-swift-storage-0\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.809081 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-db-sync-config-data\") pod \"barbican-db-sync-dnh4p\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.809857 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.822084 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-combined-ca-bundle\") pod \"barbican-db-sync-dnh4p\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.825249 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.846410 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6pz5\" (UniqueName: \"kubernetes.io/projected/4be2cd5e-1345-4212-9563-449a665090f1-kube-api-access-g6pz5\") pod \"barbican-db-sync-dnh4p\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:20 crc kubenswrapper[4862]: I1128 11:43:20.870397 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="966949c3-c79d-4811-8c75-c4e913cd6a19" path="/var/lib/kubelet/pods/966949c3-c79d-4811-8c75-c4e913cd6a19/volumes" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.906861 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7qn8\" (UniqueName: \"kubernetes.io/projected/eced1563-10aa-48f6-b79a-c017a3d4bba0-kube-api-access-n7qn8\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.906908 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.907003 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr56q\" (UniqueName: \"kubernetes.io/projected/0bca720b-962c-48e5-9861-406c78a5b202-kube-api-access-rr56q\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.907046 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-config\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.907071 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-swift-storage-0\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.907113 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-config-data\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.907145 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.907164 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-scripts\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.907180 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-combined-ca-bundle\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.907207 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eced1563-10aa-48f6-b79a-c017a3d4bba0-logs\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.907231 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-svc\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.908068 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-svc\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.909880 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-config\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.910417 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.910491 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-swift-storage-0\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.910517 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.910938 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.936676 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr56q\" (UniqueName: \"kubernetes.io/projected/0bca720b-962c-48e5-9861-406c78a5b202-kube-api-access-rr56q\") pod \"dnsmasq-dns-5f5d458b55-bfmjn\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.970062 4862 generic.go:334] "Generic (PLEG): container finished" podID="6cd448ac-1ecc-4de0-841e-08a1df0f1263" containerID="6c36a7bcb85f4756db1b4ff5b1cfb72e8f91ca6655d49fa42c182542796abcbf" exitCode=0 Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.970115 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" event={"ID":"6cd448ac-1ecc-4de0-841e-08a1df0f1263","Type":"ContainerDied","Data":"6c36a7bcb85f4756db1b4ff5b1cfb72e8f91ca6655d49fa42c182542796abcbf"} Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:20.970140 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" event={"ID":"6cd448ac-1ecc-4de0-841e-08a1df0f1263","Type":"ContainerStarted","Data":"5484cb63aa08551c497f2977d3331356464dcc4b247ae19913b1a22e713d1409"} Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.008229 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7qn8\" (UniqueName: \"kubernetes.io/projected/eced1563-10aa-48f6-b79a-c017a3d4bba0-kube-api-access-n7qn8\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.008310 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-config-data\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.008344 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-scripts\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.008362 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-combined-ca-bundle\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.008384 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eced1563-10aa-48f6-b79a-c017a3d4bba0-logs\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.008925 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eced1563-10aa-48f6-b79a-c017a3d4bba0-logs\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.019367 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.021300 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-combined-ca-bundle\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.021931 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-config-data\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.033649 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-scripts\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.033925 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7qn8\" (UniqueName: \"kubernetes.io/projected/eced1563-10aa-48f6-b79a-c017a3d4bba0-kube-api-access-n7qn8\") pod \"placement-db-sync-5lsz4\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.132382 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.297327 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.300063 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.301852 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.302031 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.302172 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.302418 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vp6qr" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.304607 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.370370 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.375054 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.378494 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.378881 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.387002 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.419019 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-logs\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.419072 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-config-data\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.419304 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.419447 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.419573 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbhxj\" (UniqueName: \"kubernetes.io/projected/f10c95ab-cbcb-4337-b07b-179915c3bac9-kube-api-access-bbhxj\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.419693 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.419729 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-scripts\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.419764 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521390 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521452 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521485 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-scripts\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521517 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521550 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-logs\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521581 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-config-data\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521610 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521651 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521673 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521712 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521742 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-logs\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521764 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521814 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521854 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68zzm\" (UniqueName: \"kubernetes.io/projected/a4822b08-b6e7-436c-b62e-d12a084e15fb-kube-api-access-68zzm\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521905 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbhxj\" (UniqueName: \"kubernetes.io/projected/f10c95ab-cbcb-4337-b07b-179915c3bac9-kube-api-access-bbhxj\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.521957 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.523125 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-logs\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.523314 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.523904 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.528614 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-config-data\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.529318 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.534717 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-scripts\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.541059 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.545553 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbhxj\" (UniqueName: \"kubernetes.io/projected/f10c95ab-cbcb-4337-b07b-179915c3bac9-kube-api-access-bbhxj\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.558274 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.623842 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68zzm\" (UniqueName: \"kubernetes.io/projected/a4822b08-b6e7-436c-b62e-d12a084e15fb-kube-api-access-68zzm\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.623927 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.623945 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.623994 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.624024 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.624067 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.624124 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-logs\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.624142 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.624451 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.624731 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.624900 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-logs\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.629262 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.629281 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.629724 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.633871 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.643141 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.646181 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68zzm\" (UniqueName: \"kubernetes.io/projected/a4822b08-b6e7-436c-b62e-d12a084e15fb-kube-api-access-68zzm\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.647458 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:21 crc kubenswrapper[4862]: I1128 11:43:21.697062 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.061261 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-654fz"] Nov 28 11:43:22 crc kubenswrapper[4862]: W1128 11:43:22.063635 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cd16bba_3339_4b8e_b495_2a8a47c6a835.slice/crio-bab3abf1e5bc7722a7b67ec1f144c8e996bfcb4c2597c402177ae2f364565739 WatchSource:0}: Error finding container bab3abf1e5bc7722a7b67ec1f144c8e996bfcb4c2597c402177ae2f364565739: Status 404 returned error can't find the container with id bab3abf1e5bc7722a7b67ec1f144c8e996bfcb4c2597c402177ae2f364565739 Nov 28 11:43:22 crc kubenswrapper[4862]: W1128 11:43:22.071837 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde671dfb_c489_4f8c_8e77_da684cca1f98.slice/crio-17f10d98d387889793f00c0533c81a3ffbd36606081c67f7a329538e7556ee35 WatchSource:0}: Error finding container 17f10d98d387889793f00c0533c81a3ffbd36606081c67f7a329538e7556ee35: Status 404 returned error can't find the container with id 17f10d98d387889793f00c0533c81a3ffbd36606081c67f7a329538e7556ee35 Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.073202 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76bb7864cf-2r6pg"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.261733 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dnh4p"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.283083 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.336516 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.444516 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.466047 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-swift-storage-0\") pod \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.466258 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-nb\") pod \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.466280 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-sb\") pod \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.466359 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-svc\") pod \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.466380 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.466446 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq8s9\" (UniqueName: \"kubernetes.io/projected/6cd448ac-1ecc-4de0-841e-08a1df0f1263-kube-api-access-bq8s9\") pod \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.466496 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-config\") pod \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\" (UID: \"6cd448ac-1ecc-4de0-841e-08a1df0f1263\") " Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.484521 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cd448ac-1ecc-4de0-841e-08a1df0f1263-kube-api-access-bq8s9" (OuterVolumeSpecName: "kube-api-access-bq8s9") pod "6cd448ac-1ecc-4de0-841e-08a1df0f1263" (UID: "6cd448ac-1ecc-4de0-841e-08a1df0f1263"). InnerVolumeSpecName "kube-api-access-bq8s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.497290 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5jcj5"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.507640 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-47sh5"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.539787 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f5d458b55-bfmjn"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.551364 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5lsz4"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.556130 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6cd448ac-1ecc-4de0-841e-08a1df0f1263" (UID: "6cd448ac-1ecc-4de0-841e-08a1df0f1263"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.558327 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6cd448ac-1ecc-4de0-841e-08a1df0f1263" (UID: "6cd448ac-1ecc-4de0-841e-08a1df0f1263"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.562873 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-config" (OuterVolumeSpecName: "config") pod "6cd448ac-1ecc-4de0-841e-08a1df0f1263" (UID: "6cd448ac-1ecc-4de0-841e-08a1df0f1263"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.572352 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.572387 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.572396 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq8s9\" (UniqueName: \"kubernetes.io/projected/6cd448ac-1ecc-4de0-841e-08a1df0f1263-kube-api-access-bq8s9\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.572407 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.576890 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.584441 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6cd448ac-1ecc-4de0-841e-08a1df0f1263" (UID: "6cd448ac-1ecc-4de0-841e-08a1df0f1263"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:22 crc kubenswrapper[4862]: W1128 11:43:22.605465 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeced1563_10aa_48f6_b79a_c017a3d4bba0.slice/crio-3baf996dbc637e47dd7542ac9960d71202c30cdbd3949613892fec7f6ae657f4 WatchSource:0}: Error finding container 3baf996dbc637e47dd7542ac9960d71202c30cdbd3949613892fec7f6ae657f4: Status 404 returned error can't find the container with id 3baf996dbc637e47dd7542ac9960d71202c30cdbd3949613892fec7f6ae657f4 Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.609974 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6cd448ac-1ecc-4de0-841e-08a1df0f1263" (UID: "6cd448ac-1ecc-4de0-841e-08a1df0f1263"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.648215 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.673823 4862 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.675661 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cd448ac-1ecc-4de0-841e-08a1df0f1263-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:22 crc kubenswrapper[4862]: I1128 11:43:22.786239 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.030074 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-47sh5" event={"ID":"82ea93c1-317d-4d0f-92a6-93872ef50d12","Type":"ContainerStarted","Data":"e9feb955f32a4be56fdec8757fa240832816eb7bcbb5e3c93bb365525206777c"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.037527 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5jcj5" event={"ID":"b376a4e9-5456-4e8f-9c91-b222f38c30db","Type":"ContainerStarted","Data":"33eb29d50e1b0895601a8eb8990f3270f01e8748228dbc7f014a85b73419dda1"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.042752 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-654fz" event={"ID":"9cd16bba-3339-4b8e-b495-2a8a47c6a835","Type":"ContainerStarted","Data":"57a230e1e4593a57f1af78d10623daad185b54c29ca745c19aa3ecc84aec7fbe"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.042801 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-654fz" event={"ID":"9cd16bba-3339-4b8e-b495-2a8a47c6a835","Type":"ContainerStarted","Data":"bab3abf1e5bc7722a7b67ec1f144c8e996bfcb4c2597c402177ae2f364565739"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.045691 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a4822b08-b6e7-436c-b62e-d12a084e15fb","Type":"ContainerStarted","Data":"7b63b58ba476bc6172b81e3717b827aff3f2eaae4f0aeaedb8571c61616ed393"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.048973 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5lsz4" event={"ID":"eced1563-10aa-48f6-b79a-c017a3d4bba0","Type":"ContainerStarted","Data":"3baf996dbc637e47dd7542ac9960d71202c30cdbd3949613892fec7f6ae657f4"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.057413 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f10c95ab-cbcb-4337-b07b-179915c3bac9","Type":"ContainerStarted","Data":"1062463765b62df31d0aa87cf66570f646ea9b0b09e73ce29c7012447246a9b8"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.060630 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dnh4p" event={"ID":"4be2cd5e-1345-4212-9563-449a665090f1","Type":"ContainerStarted","Data":"1e18d482ffa1b6d12eaece1875d3baa5e50a3d241ba5e9fe08357138da42398f"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.065789 4862 generic.go:334] "Generic (PLEG): container finished" podID="de671dfb-c489-4f8c-8e77-da684cca1f98" containerID="341b5408e8fba76c898493fbba0262347e5f02ebc2fee5b91b351dc3545f76f1" exitCode=0 Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.065888 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" event={"ID":"de671dfb-c489-4f8c-8e77-da684cca1f98","Type":"ContainerDied","Data":"341b5408e8fba76c898493fbba0262347e5f02ebc2fee5b91b351dc3545f76f1"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.065997 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" event={"ID":"de671dfb-c489-4f8c-8e77-da684cca1f98","Type":"ContainerStarted","Data":"17f10d98d387889793f00c0533c81a3ffbd36606081c67f7a329538e7556ee35"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.070257 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" event={"ID":"0bca720b-962c-48e5-9861-406c78a5b202","Type":"ContainerStarted","Data":"dc4975e4749b7934bc65feb740a704378b7be1920f999976219d27caaa9eae16"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.075210 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerStarted","Data":"5dc6b2417038674a878c4d4565581172dd45cbff0efcbb6b441236e24c798b9d"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.076846 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" event={"ID":"6cd448ac-1ecc-4de0-841e-08a1df0f1263","Type":"ContainerDied","Data":"5484cb63aa08551c497f2977d3331356464dcc4b247ae19913b1a22e713d1409"} Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.076910 4862 scope.go:117] "RemoveContainer" containerID="6c36a7bcb85f4756db1b4ff5b1cfb72e8f91ca6655d49fa42c182542796abcbf" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.076935 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6bbf7467-wd82l" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.139185 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6bbf7467-wd82l"] Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.147890 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b6bbf7467-wd82l"] Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.152788 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-654fz" podStartSLOduration=3.152767129 podStartE2EDuration="3.152767129s" podCreationTimestamp="2025-11-28 11:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:23.151433684 +0000 UTC m=+1280.683947595" watchObservedRunningTime="2025-11-28 11:43:23.152767129 +0000 UTC m=+1280.685281050" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.449149 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.511569 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-config\") pod \"de671dfb-c489-4f8c-8e77-da684cca1f98\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.511653 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-sb\") pod \"de671dfb-c489-4f8c-8e77-da684cca1f98\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.511710 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2swg\" (UniqueName: \"kubernetes.io/projected/de671dfb-c489-4f8c-8e77-da684cca1f98-kube-api-access-l2swg\") pod \"de671dfb-c489-4f8c-8e77-da684cca1f98\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.511792 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-nb\") pod \"de671dfb-c489-4f8c-8e77-da684cca1f98\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.511823 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-swift-storage-0\") pod \"de671dfb-c489-4f8c-8e77-da684cca1f98\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.512000 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-svc\") pod \"de671dfb-c489-4f8c-8e77-da684cca1f98\" (UID: \"de671dfb-c489-4f8c-8e77-da684cca1f98\") " Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.519780 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de671dfb-c489-4f8c-8e77-da684cca1f98-kube-api-access-l2swg" (OuterVolumeSpecName: "kube-api-access-l2swg") pod "de671dfb-c489-4f8c-8e77-da684cca1f98" (UID: "de671dfb-c489-4f8c-8e77-da684cca1f98"). InnerVolumeSpecName "kube-api-access-l2swg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.542836 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "de671dfb-c489-4f8c-8e77-da684cca1f98" (UID: "de671dfb-c489-4f8c-8e77-da684cca1f98"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.553792 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-config" (OuterVolumeSpecName: "config") pod "de671dfb-c489-4f8c-8e77-da684cca1f98" (UID: "de671dfb-c489-4f8c-8e77-da684cca1f98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.559034 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de671dfb-c489-4f8c-8e77-da684cca1f98" (UID: "de671dfb-c489-4f8c-8e77-da684cca1f98"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.564972 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de671dfb-c489-4f8c-8e77-da684cca1f98" (UID: "de671dfb-c489-4f8c-8e77-da684cca1f98"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.573574 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de671dfb-c489-4f8c-8e77-da684cca1f98" (UID: "de671dfb-c489-4f8c-8e77-da684cca1f98"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.615205 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.615253 4862 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.615267 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.615311 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.615323 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de671dfb-c489-4f8c-8e77-da684cca1f98-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:23 crc kubenswrapper[4862]: I1128 11:43:23.615335 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2swg\" (UniqueName: \"kubernetes.io/projected/de671dfb-c489-4f8c-8e77-da684cca1f98-kube-api-access-l2swg\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.115316 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-47sh5" event={"ID":"82ea93c1-317d-4d0f-92a6-93872ef50d12","Type":"ContainerStarted","Data":"cf25724267080641f4ff63f79418c992c069ccd67889b4a80b2782cf3e9a2d04"} Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.121027 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a4822b08-b6e7-436c-b62e-d12a084e15fb","Type":"ContainerStarted","Data":"45ccdccdf54f084bcddacce3e741ceac3305ffa65525771df4d633cb9ff5a1f6"} Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.132387 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" event={"ID":"de671dfb-c489-4f8c-8e77-da684cca1f98","Type":"ContainerDied","Data":"17f10d98d387889793f00c0533c81a3ffbd36606081c67f7a329538e7556ee35"} Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.132446 4862 scope.go:117] "RemoveContainer" containerID="341b5408e8fba76c898493fbba0262347e5f02ebc2fee5b91b351dc3545f76f1" Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.132573 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76bb7864cf-2r6pg" Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.132874 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-47sh5" podStartSLOduration=4.132864541 podStartE2EDuration="4.132864541s" podCreationTimestamp="2025-11-28 11:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:24.129231995 +0000 UTC m=+1281.661745916" watchObservedRunningTime="2025-11-28 11:43:24.132864541 +0000 UTC m=+1281.665378462" Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.140486 4862 generic.go:334] "Generic (PLEG): container finished" podID="0bca720b-962c-48e5-9861-406c78a5b202" containerID="18e893787b962c2c301c113217c240503caa6893de81f9152dd44079187812f3" exitCode=0 Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.140564 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" event={"ID":"0bca720b-962c-48e5-9861-406c78a5b202","Type":"ContainerDied","Data":"18e893787b962c2c301c113217c240503caa6893de81f9152dd44079187812f3"} Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.145000 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f10c95ab-cbcb-4337-b07b-179915c3bac9","Type":"ContainerStarted","Data":"bf7622e5758d9c528deb5913d5a39d809ac5c43c954e697129005ed0d90fc6a5"} Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.240618 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76bb7864cf-2r6pg"] Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.242682 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76bb7864cf-2r6pg"] Nov 28 11:43:24 crc kubenswrapper[4862]: I1128 11:43:24.873722 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cd448ac-1ecc-4de0-841e-08a1df0f1263" path="/var/lib/kubelet/pods/6cd448ac-1ecc-4de0-841e-08a1df0f1263/volumes" Nov 28 11:43:25 crc kubenswrapper[4862]: I1128 11:43:25.120737 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de671dfb-c489-4f8c-8e77-da684cca1f98" path="/var/lib/kubelet/pods/de671dfb-c489-4f8c-8e77-da684cca1f98/volumes" Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.171376 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" event={"ID":"0bca720b-962c-48e5-9861-406c78a5b202","Type":"ContainerStarted","Data":"b570c3eb66bb4f0f229dea23e06f021c255107f84372e2683901f74f1fa171e4"} Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.171778 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.180294 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f10c95ab-cbcb-4337-b07b-179915c3bac9","Type":"ContainerStarted","Data":"adf70a4ef26a503fbddaba6b78035a1836fd47d768c709e4b11b7bb63c532881"} Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.180441 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerName="glance-log" containerID="cri-o://bf7622e5758d9c528deb5913d5a39d809ac5c43c954e697129005ed0d90fc6a5" gracePeriod=30 Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.180742 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerName="glance-httpd" containerID="cri-o://adf70a4ef26a503fbddaba6b78035a1836fd47d768c709e4b11b7bb63c532881" gracePeriod=30 Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.193721 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a4822b08-b6e7-436c-b62e-d12a084e15fb","Type":"ContainerStarted","Data":"1b8c69321e5670cb29b365186ef46a7d4583672b5b3573aa8946231442f89179"} Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.194062 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerName="glance-log" containerID="cri-o://45ccdccdf54f084bcddacce3e741ceac3305ffa65525771df4d633cb9ff5a1f6" gracePeriod=30 Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.194267 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerName="glance-httpd" containerID="cri-o://1b8c69321e5670cb29b365186ef46a7d4583672b5b3573aa8946231442f89179" gracePeriod=30 Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.211247 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" podStartSLOduration=6.211224702 podStartE2EDuration="6.211224702s" podCreationTimestamp="2025-11-28 11:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:26.196507071 +0000 UTC m=+1283.729020992" watchObservedRunningTime="2025-11-28 11:43:26.211224702 +0000 UTC m=+1283.743738633" Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.259378 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.25935609 podStartE2EDuration="6.25935609s" podCreationTimestamp="2025-11-28 11:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:26.240320384 +0000 UTC m=+1283.772834355" watchObservedRunningTime="2025-11-28 11:43:26.25935609 +0000 UTC m=+1283.791870011" Nov 28 11:43:26 crc kubenswrapper[4862]: I1128 11:43:26.268359 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.268336918 podStartE2EDuration="6.268336918s" podCreationTimestamp="2025-11-28 11:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:26.262434781 +0000 UTC m=+1283.794948702" watchObservedRunningTime="2025-11-28 11:43:26.268336918 +0000 UTC m=+1283.800850839" Nov 28 11:43:27 crc kubenswrapper[4862]: I1128 11:43:27.204352 4862 generic.go:334] "Generic (PLEG): container finished" podID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerID="adf70a4ef26a503fbddaba6b78035a1836fd47d768c709e4b11b7bb63c532881" exitCode=0 Nov 28 11:43:27 crc kubenswrapper[4862]: I1128 11:43:27.204620 4862 generic.go:334] "Generic (PLEG): container finished" podID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerID="bf7622e5758d9c528deb5913d5a39d809ac5c43c954e697129005ed0d90fc6a5" exitCode=143 Nov 28 11:43:27 crc kubenswrapper[4862]: I1128 11:43:27.204427 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f10c95ab-cbcb-4337-b07b-179915c3bac9","Type":"ContainerDied","Data":"adf70a4ef26a503fbddaba6b78035a1836fd47d768c709e4b11b7bb63c532881"} Nov 28 11:43:27 crc kubenswrapper[4862]: I1128 11:43:27.204692 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f10c95ab-cbcb-4337-b07b-179915c3bac9","Type":"ContainerDied","Data":"bf7622e5758d9c528deb5913d5a39d809ac5c43c954e697129005ed0d90fc6a5"} Nov 28 11:43:27 crc kubenswrapper[4862]: I1128 11:43:27.207624 4862 generic.go:334] "Generic (PLEG): container finished" podID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerID="1b8c69321e5670cb29b365186ef46a7d4583672b5b3573aa8946231442f89179" exitCode=0 Nov 28 11:43:27 crc kubenswrapper[4862]: I1128 11:43:27.207645 4862 generic.go:334] "Generic (PLEG): container finished" podID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerID="45ccdccdf54f084bcddacce3e741ceac3305ffa65525771df4d633cb9ff5a1f6" exitCode=143 Nov 28 11:43:27 crc kubenswrapper[4862]: I1128 11:43:27.207854 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a4822b08-b6e7-436c-b62e-d12a084e15fb","Type":"ContainerDied","Data":"1b8c69321e5670cb29b365186ef46a7d4583672b5b3573aa8946231442f89179"} Nov 28 11:43:27 crc kubenswrapper[4862]: I1128 11:43:27.207900 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a4822b08-b6e7-436c-b62e-d12a084e15fb","Type":"ContainerDied","Data":"45ccdccdf54f084bcddacce3e741ceac3305ffa65525771df4d633cb9ff5a1f6"} Nov 28 11:43:28 crc kubenswrapper[4862]: I1128 11:43:28.234791 4862 generic.go:334] "Generic (PLEG): container finished" podID="9cd16bba-3339-4b8e-b495-2a8a47c6a835" containerID="57a230e1e4593a57f1af78d10623daad185b54c29ca745c19aa3ecc84aec7fbe" exitCode=0 Nov 28 11:43:28 crc kubenswrapper[4862]: I1128 11:43:28.235180 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-654fz" event={"ID":"9cd16bba-3339-4b8e-b495-2a8a47c6a835","Type":"ContainerDied","Data":"57a230e1e4593a57f1af78d10623daad185b54c29ca745c19aa3ecc84aec7fbe"} Nov 28 11:43:31 crc kubenswrapper[4862]: I1128 11:43:31.022288 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:43:31 crc kubenswrapper[4862]: I1128 11:43:31.097408 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dd4845ddf-7ngnt"] Nov 28 11:43:31 crc kubenswrapper[4862]: I1128 11:43:31.097798 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerName="dnsmasq-dns" containerID="cri-o://d16ba579580174081c4341491a21bdbf0d1f32acd583f99f26471ced2f0fe306" gracePeriod=10 Nov 28 11:43:32 crc kubenswrapper[4862]: I1128 11:43:32.305840 4862 generic.go:334] "Generic (PLEG): container finished" podID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerID="d16ba579580174081c4341491a21bdbf0d1f32acd583f99f26471ced2f0fe306" exitCode=0 Nov 28 11:43:32 crc kubenswrapper[4862]: I1128 11:43:32.305906 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" event={"ID":"c657ab4a-f20b-46a1-aa1b-35aed553d6c0","Type":"ContainerDied","Data":"d16ba579580174081c4341491a21bdbf0d1f32acd583f99f26471ced2f0fe306"} Nov 28 11:43:33 crc kubenswrapper[4862]: I1128 11:43:33.273838 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.162664 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.164519 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.321833 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-scripts\") pod \"f10c95ab-cbcb-4337-b07b-179915c3bac9\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.321887 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-credential-keys\") pod \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.321924 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-scripts\") pod \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.321952 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-fernet-keys\") pod \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.321989 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-logs\") pod \"f10c95ab-cbcb-4337-b07b-179915c3bac9\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.322045 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-config-data\") pod \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.322065 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbhxj\" (UniqueName: \"kubernetes.io/projected/f10c95ab-cbcb-4337-b07b-179915c3bac9-kube-api-access-bbhxj\") pod \"f10c95ab-cbcb-4337-b07b-179915c3bac9\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.322113 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-public-tls-certs\") pod \"f10c95ab-cbcb-4337-b07b-179915c3bac9\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.322133 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-config-data\") pod \"f10c95ab-cbcb-4337-b07b-179915c3bac9\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.322147 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"f10c95ab-cbcb-4337-b07b-179915c3bac9\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.322192 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl57w\" (UniqueName: \"kubernetes.io/projected/9cd16bba-3339-4b8e-b495-2a8a47c6a835-kube-api-access-kl57w\") pod \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.322213 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-combined-ca-bundle\") pod \"f10c95ab-cbcb-4337-b07b-179915c3bac9\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.322288 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-combined-ca-bundle\") pod \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\" (UID: \"9cd16bba-3339-4b8e-b495-2a8a47c6a835\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.322320 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-httpd-run\") pod \"f10c95ab-cbcb-4337-b07b-179915c3bac9\" (UID: \"f10c95ab-cbcb-4337-b07b-179915c3bac9\") " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.323957 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f10c95ab-cbcb-4337-b07b-179915c3bac9" (UID: "f10c95ab-cbcb-4337-b07b-179915c3bac9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.327317 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-logs" (OuterVolumeSpecName: "logs") pod "f10c95ab-cbcb-4337-b07b-179915c3bac9" (UID: "f10c95ab-cbcb-4337-b07b-179915c3bac9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.329647 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-scripts" (OuterVolumeSpecName: "scripts") pod "9cd16bba-3339-4b8e-b495-2a8a47c6a835" (UID: "9cd16bba-3339-4b8e-b495-2a8a47c6a835"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.330224 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "f10c95ab-cbcb-4337-b07b-179915c3bac9" (UID: "f10c95ab-cbcb-4337-b07b-179915c3bac9"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.331365 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f10c95ab-cbcb-4337-b07b-179915c3bac9-kube-api-access-bbhxj" (OuterVolumeSpecName: "kube-api-access-bbhxj") pod "f10c95ab-cbcb-4337-b07b-179915c3bac9" (UID: "f10c95ab-cbcb-4337-b07b-179915c3bac9"). InnerVolumeSpecName "kube-api-access-bbhxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.332141 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd16bba-3339-4b8e-b495-2a8a47c6a835-kube-api-access-kl57w" (OuterVolumeSpecName: "kube-api-access-kl57w") pod "9cd16bba-3339-4b8e-b495-2a8a47c6a835" (UID: "9cd16bba-3339-4b8e-b495-2a8a47c6a835"). InnerVolumeSpecName "kube-api-access-kl57w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.332735 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-scripts" (OuterVolumeSpecName: "scripts") pod "f10c95ab-cbcb-4337-b07b-179915c3bac9" (UID: "f10c95ab-cbcb-4337-b07b-179915c3bac9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.333022 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9cd16bba-3339-4b8e-b495-2a8a47c6a835" (UID: "9cd16bba-3339-4b8e-b495-2a8a47c6a835"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.335427 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9cd16bba-3339-4b8e-b495-2a8a47c6a835" (UID: "9cd16bba-3339-4b8e-b495-2a8a47c6a835"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.349489 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f10c95ab-cbcb-4337-b07b-179915c3bac9","Type":"ContainerDied","Data":"1062463765b62df31d0aa87cf66570f646ea9b0b09e73ce29c7012447246a9b8"} Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.349548 4862 scope.go:117] "RemoveContainer" containerID="adf70a4ef26a503fbddaba6b78035a1836fd47d768c709e4b11b7bb63c532881" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.349979 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.354167 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-654fz" event={"ID":"9cd16bba-3339-4b8e-b495-2a8a47c6a835","Type":"ContainerDied","Data":"bab3abf1e5bc7722a7b67ec1f144c8e996bfcb4c2597c402177ae2f364565739"} Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.354250 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bab3abf1e5bc7722a7b67ec1f144c8e996bfcb4c2597c402177ae2f364565739" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.354347 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-654fz" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.359681 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-config-data" (OuterVolumeSpecName: "config-data") pod "9cd16bba-3339-4b8e-b495-2a8a47c6a835" (UID: "9cd16bba-3339-4b8e-b495-2a8a47c6a835"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.378508 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cd16bba-3339-4b8e-b495-2a8a47c6a835" (UID: "9cd16bba-3339-4b8e-b495-2a8a47c6a835"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.384219 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f10c95ab-cbcb-4337-b07b-179915c3bac9" (UID: "f10c95ab-cbcb-4337-b07b-179915c3bac9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.385516 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f10c95ab-cbcb-4337-b07b-179915c3bac9" (UID: "f10c95ab-cbcb-4337-b07b-179915c3bac9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.409346 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-config-data" (OuterVolumeSpecName: "config-data") pod "f10c95ab-cbcb-4337-b07b-179915c3bac9" (UID: "f10c95ab-cbcb-4337-b07b-179915c3bac9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425618 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425664 4862 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425681 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425693 4862 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425705 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425716 4862 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425728 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10c95ab-cbcb-4337-b07b-179915c3bac9-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425739 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cd16bba-3339-4b8e-b495-2a8a47c6a835-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425750 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbhxj\" (UniqueName: \"kubernetes.io/projected/f10c95ab-cbcb-4337-b07b-179915c3bac9-kube-api-access-bbhxj\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425762 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425775 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425816 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425828 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl57w\" (UniqueName: \"kubernetes.io/projected/9cd16bba-3339-4b8e-b495-2a8a47c6a835-kube-api-access-kl57w\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.425839 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10c95ab-cbcb-4337-b07b-179915c3bac9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.453326 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.528968 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.712685 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.719838 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.747438 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:43:35 crc kubenswrapper[4862]: E1128 11:43:35.747799 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de671dfb-c489-4f8c-8e77-da684cca1f98" containerName="init" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.747815 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="de671dfb-c489-4f8c-8e77-da684cca1f98" containerName="init" Nov 28 11:43:35 crc kubenswrapper[4862]: E1128 11:43:35.747839 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd448ac-1ecc-4de0-841e-08a1df0f1263" containerName="init" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.747845 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd448ac-1ecc-4de0-841e-08a1df0f1263" containerName="init" Nov 28 11:43:35 crc kubenswrapper[4862]: E1128 11:43:35.747855 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd16bba-3339-4b8e-b495-2a8a47c6a835" containerName="keystone-bootstrap" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.747861 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd16bba-3339-4b8e-b495-2a8a47c6a835" containerName="keystone-bootstrap" Nov 28 11:43:35 crc kubenswrapper[4862]: E1128 11:43:35.747876 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerName="glance-httpd" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.747883 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerName="glance-httpd" Nov 28 11:43:35 crc kubenswrapper[4862]: E1128 11:43:35.747895 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerName="glance-log" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.747901 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerName="glance-log" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.748086 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd16bba-3339-4b8e-b495-2a8a47c6a835" containerName="keystone-bootstrap" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.748116 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="de671dfb-c489-4f8c-8e77-da684cca1f98" containerName="init" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.748126 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerName="glance-httpd" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.748136 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd448ac-1ecc-4de0-841e-08a1df0f1263" containerName="init" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.748144 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10c95ab-cbcb-4337-b07b-179915c3bac9" containerName="glance-log" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.748942 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.751171 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.751612 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.766316 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.938048 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.938138 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.938413 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-config-data\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.938520 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.938551 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-scripts\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.938605 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ckmp\" (UniqueName: \"kubernetes.io/projected/72c062a5-78c7-407a-8c0d-858a6fc301e2-kube-api-access-8ckmp\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.938636 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:35 crc kubenswrapper[4862]: I1128 11:43:35.938656 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-logs\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.040992 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041071 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-scripts\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041116 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ckmp\" (UniqueName: \"kubernetes.io/projected/72c062a5-78c7-407a-8c0d-858a6fc301e2-kube-api-access-8ckmp\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041147 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041174 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-logs\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041245 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041279 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041363 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-config-data\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041529 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041784 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-logs\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.041895 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.046293 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-scripts\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.046983 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-config-data\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.049974 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.050892 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.058341 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ckmp\" (UniqueName: \"kubernetes.io/projected/72c062a5-78c7-407a-8c0d-858a6fc301e2-kube-api-access-8ckmp\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.070736 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.368554 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.410313 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-654fz"] Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.429195 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-654fz"] Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.532007 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-527fq"] Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.533746 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.536919 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.536960 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.537186 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gn5r2" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.537356 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.537472 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.579348 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-527fq"] Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.655485 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-config-data\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.655586 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-scripts\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.655646 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-combined-ca-bundle\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.655672 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-fernet-keys\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.655698 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-credential-keys\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.655794 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk7q4\" (UniqueName: \"kubernetes.io/projected/57dc2822-b0b6-45d4-92de-5f601ebcf48a-kube-api-access-pk7q4\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.757234 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk7q4\" (UniqueName: \"kubernetes.io/projected/57dc2822-b0b6-45d4-92de-5f601ebcf48a-kube-api-access-pk7q4\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.757314 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-config-data\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.757379 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-scripts\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.757431 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-combined-ca-bundle\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.757455 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-fernet-keys\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.757481 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-credential-keys\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.762320 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-credential-keys\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.763085 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-config-data\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.763232 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-combined-ca-bundle\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.763395 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-fernet-keys\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.768375 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-scripts\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.788924 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk7q4\" (UniqueName: \"kubernetes.io/projected/57dc2822-b0b6-45d4-92de-5f601ebcf48a-kube-api-access-pk7q4\") pod \"keystone-bootstrap-527fq\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.854912 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.855163 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cd16bba-3339-4b8e-b495-2a8a47c6a835" path="/var/lib/kubelet/pods/9cd16bba-3339-4b8e-b495-2a8a47c6a835/volumes" Nov 28 11:43:36 crc kubenswrapper[4862]: I1128 11:43:36.856812 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f10c95ab-cbcb-4337-b07b-179915c3bac9" path="/var/lib/kubelet/pods/f10c95ab-cbcb-4337-b07b-179915c3bac9/volumes" Nov 28 11:43:43 crc kubenswrapper[4862]: I1128 11:43:43.273621 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Nov 28 11:43:43 crc kubenswrapper[4862]: I1128 11:43:43.949006 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.006226 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-internal-tls-certs\") pod \"a4822b08-b6e7-436c-b62e-d12a084e15fb\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.006315 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68zzm\" (UniqueName: \"kubernetes.io/projected/a4822b08-b6e7-436c-b62e-d12a084e15fb-kube-api-access-68zzm\") pod \"a4822b08-b6e7-436c-b62e-d12a084e15fb\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.006399 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-logs\") pod \"a4822b08-b6e7-436c-b62e-d12a084e15fb\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.006493 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"a4822b08-b6e7-436c-b62e-d12a084e15fb\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.006595 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-combined-ca-bundle\") pod \"a4822b08-b6e7-436c-b62e-d12a084e15fb\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.006745 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-config-data\") pod \"a4822b08-b6e7-436c-b62e-d12a084e15fb\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.006824 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-httpd-run\") pod \"a4822b08-b6e7-436c-b62e-d12a084e15fb\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.006933 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-scripts\") pod \"a4822b08-b6e7-436c-b62e-d12a084e15fb\" (UID: \"a4822b08-b6e7-436c-b62e-d12a084e15fb\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.007608 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a4822b08-b6e7-436c-b62e-d12a084e15fb" (UID: "a4822b08-b6e7-436c-b62e-d12a084e15fb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.007771 4862 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.007490 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-logs" (OuterVolumeSpecName: "logs") pod "a4822b08-b6e7-436c-b62e-d12a084e15fb" (UID: "a4822b08-b6e7-436c-b62e-d12a084e15fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.014189 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "a4822b08-b6e7-436c-b62e-d12a084e15fb" (UID: "a4822b08-b6e7-436c-b62e-d12a084e15fb"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.014756 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4822b08-b6e7-436c-b62e-d12a084e15fb-kube-api-access-68zzm" (OuterVolumeSpecName: "kube-api-access-68zzm") pod "a4822b08-b6e7-436c-b62e-d12a084e15fb" (UID: "a4822b08-b6e7-436c-b62e-d12a084e15fb"). InnerVolumeSpecName "kube-api-access-68zzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.014946 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-scripts" (OuterVolumeSpecName: "scripts") pod "a4822b08-b6e7-436c-b62e-d12a084e15fb" (UID: "a4822b08-b6e7-436c-b62e-d12a084e15fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.062617 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4822b08-b6e7-436c-b62e-d12a084e15fb" (UID: "a4822b08-b6e7-436c-b62e-d12a084e15fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.075400 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-config-data" (OuterVolumeSpecName: "config-data") pod "a4822b08-b6e7-436c-b62e-d12a084e15fb" (UID: "a4822b08-b6e7-436c-b62e-d12a084e15fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.109653 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68zzm\" (UniqueName: \"kubernetes.io/projected/a4822b08-b6e7-436c-b62e-d12a084e15fb-kube-api-access-68zzm\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.109686 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4822b08-b6e7-436c-b62e-d12a084e15fb-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.109714 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.109725 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.109734 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.109741 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.115721 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a4822b08-b6e7-436c-b62e-d12a084e15fb" (UID: "a4822b08-b6e7-436c-b62e-d12a084e15fb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.129171 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.212121 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4822b08-b6e7-436c-b62e-d12a084e15fb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.212405 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.459482 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a4822b08-b6e7-436c-b62e-d12a084e15fb","Type":"ContainerDied","Data":"7b63b58ba476bc6172b81e3717b827aff3f2eaae4f0aeaedb8571c61616ed393"} Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.459564 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.502041 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.525861 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.547698 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:43:44 crc kubenswrapper[4862]: E1128 11:43:44.549370 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerName="glance-httpd" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.549409 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerName="glance-httpd" Nov 28 11:43:44 crc kubenswrapper[4862]: E1128 11:43:44.549431 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerName="glance-log" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.549438 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerName="glance-log" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.549765 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerName="glance-httpd" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.549794 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4822b08-b6e7-436c-b62e-d12a084e15fb" containerName="glance-log" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.551377 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.560237 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.560823 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.572988 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.627232 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.627288 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-logs\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.627319 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wc7f\" (UniqueName: \"kubernetes.io/projected/635693e3-57cb-479e-b4b0-947f81f10e97-kube-api-access-5wc7f\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.627424 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.627489 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.627531 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-scripts\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.628266 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.628378 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-config-data\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.664245 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:44 crc kubenswrapper[4862]: E1128 11:43:44.667349 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:3a56b50437a0c9a9a7b30c10f5e43bbdb7d9a94b723c70d36f0b01ff545e00eb" Nov 28 11:43:44 crc kubenswrapper[4862]: E1128 11:43:44.667655 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:3a56b50437a0c9a9a7b30c10f5e43bbdb7d9a94b723c70d36f0b01ff545e00eb,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g6pz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-dnh4p_openstack(4be2cd5e-1345-4212-9563-449a665090f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:43:44 crc kubenswrapper[4862]: E1128 11:43:44.669765 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-dnh4p" podUID="4be2cd5e-1345-4212-9563-449a665090f1" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.729950 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-nb\") pod \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730017 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-sb\") pod \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730051 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-config\") pod \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730175 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-dns-svc\") pod \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730324 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7fbj\" (UniqueName: \"kubernetes.io/projected/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-kube-api-access-q7fbj\") pod \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\" (UID: \"c657ab4a-f20b-46a1-aa1b-35aed553d6c0\") " Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730538 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-scripts\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730718 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730769 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-config-data\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730796 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730818 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-logs\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730850 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wc7f\" (UniqueName: \"kubernetes.io/projected/635693e3-57cb-479e-b4b0-947f81f10e97-kube-api-access-5wc7f\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730874 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.730903 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.731445 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.731813 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-logs\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.732377 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.737597 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-config-data\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.737765 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.738985 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.740300 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-kube-api-access-q7fbj" (OuterVolumeSpecName: "kube-api-access-q7fbj") pod "c657ab4a-f20b-46a1-aa1b-35aed553d6c0" (UID: "c657ab4a-f20b-46a1-aa1b-35aed553d6c0"). InnerVolumeSpecName "kube-api-access-q7fbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.747378 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-scripts\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.757687 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wc7f\" (UniqueName: \"kubernetes.io/projected/635693e3-57cb-479e-b4b0-947f81f10e97-kube-api-access-5wc7f\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.811560 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.819203 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-config" (OuterVolumeSpecName: "config") pod "c657ab4a-f20b-46a1-aa1b-35aed553d6c0" (UID: "c657ab4a-f20b-46a1-aa1b-35aed553d6c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.821507 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c657ab4a-f20b-46a1-aa1b-35aed553d6c0" (UID: "c657ab4a-f20b-46a1-aa1b-35aed553d6c0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.833957 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7fbj\" (UniqueName: \"kubernetes.io/projected/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-kube-api-access-q7fbj\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.834236 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.834331 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.874667 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c657ab4a-f20b-46a1-aa1b-35aed553d6c0" (UID: "c657ab4a-f20b-46a1-aa1b-35aed553d6c0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.895331 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.909159 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4822b08-b6e7-436c-b62e-d12a084e15fb" path="/var/lib/kubelet/pods/a4822b08-b6e7-436c-b62e-d12a084e15fb/volumes" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.938199 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:44 crc kubenswrapper[4862]: I1128 11:43:44.991572 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c657ab4a-f20b-46a1-aa1b-35aed553d6c0" (UID: "c657ab4a-f20b-46a1-aa1b-35aed553d6c0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:43:45 crc kubenswrapper[4862]: I1128 11:43:45.040001 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c657ab4a-f20b-46a1-aa1b-35aed553d6c0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:45 crc kubenswrapper[4862]: I1128 11:43:45.472255 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" event={"ID":"c657ab4a-f20b-46a1-aa1b-35aed553d6c0","Type":"ContainerDied","Data":"d6de6a66a66e33666b4b81bcb0a6452ef09947628ef5a084c0f836a4f903e2d6"} Nov 28 11:43:45 crc kubenswrapper[4862]: I1128 11:43:45.472273 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" Nov 28 11:43:45 crc kubenswrapper[4862]: E1128 11:43:45.476169 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:3a56b50437a0c9a9a7b30c10f5e43bbdb7d9a94b723c70d36f0b01ff545e00eb\\\"\"" pod="openstack/barbican-db-sync-dnh4p" podUID="4be2cd5e-1345-4212-9563-449a665090f1" Nov 28 11:43:45 crc kubenswrapper[4862]: I1128 11:43:45.513460 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dd4845ddf-7ngnt"] Nov 28 11:43:45 crc kubenswrapper[4862]: I1128 11:43:45.520220 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7dd4845ddf-7ngnt"] Nov 28 11:43:46 crc kubenswrapper[4862]: I1128 11:43:46.856576 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" path="/var/lib/kubelet/pods/c657ab4a-f20b-46a1-aa1b-35aed553d6c0/volumes" Nov 28 11:43:48 crc kubenswrapper[4862]: I1128 11:43:48.274521 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7dd4845ddf-7ngnt" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Nov 28 11:43:48 crc kubenswrapper[4862]: I1128 11:43:48.532477 4862 generic.go:334] "Generic (PLEG): container finished" podID="82ea93c1-317d-4d0f-92a6-93872ef50d12" containerID="cf25724267080641f4ff63f79418c992c069ccd67889b4a80b2782cf3e9a2d04" exitCode=0 Nov 28 11:43:48 crc kubenswrapper[4862]: I1128 11:43:48.532546 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-47sh5" event={"ID":"82ea93c1-317d-4d0f-92a6-93872ef50d12","Type":"ContainerDied","Data":"cf25724267080641f4ff63f79418c992c069ccd67889b4a80b2782cf3e9a2d04"} Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.346584 4862 scope.go:117] "RemoveContainer" containerID="bf7622e5758d9c528deb5913d5a39d809ac5c43c954e697129005ed0d90fc6a5" Nov 28 11:43:50 crc kubenswrapper[4862]: E1128 11:43:50.376471 4862 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b5266c9a26766fce2b92f95dff52d362a760f7baf1474cdcb33bd68570e096c0" Nov 28 11:43:50 crc kubenswrapper[4862]: E1128 11:43:50.376646 4862 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b5266c9a26766fce2b92f95dff52d362a760f7baf1474cdcb33bd68570e096c0,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5xdmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-5jcj5_openstack(b376a4e9-5456-4e8f-9c91-b222f38c30db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 11:43:50 crc kubenswrapper[4862]: E1128 11:43:50.378122 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-5jcj5" podUID="b376a4e9-5456-4e8f-9c91-b222f38c30db" Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.486230 4862 scope.go:117] "RemoveContainer" containerID="1b8c69321e5670cb29b365186ef46a7d4583672b5b3573aa8946231442f89179" Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.560688 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-47sh5" event={"ID":"82ea93c1-317d-4d0f-92a6-93872ef50d12","Type":"ContainerDied","Data":"e9feb955f32a4be56fdec8757fa240832816eb7bcbb5e3c93bb365525206777c"} Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.560733 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9feb955f32a4be56fdec8757fa240832816eb7bcbb5e3c93bb365525206777c" Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.597571 4862 scope.go:117] "RemoveContainer" containerID="45ccdccdf54f084bcddacce3e741ceac3305ffa65525771df4d633cb9ff5a1f6" Nov 28 11:43:50 crc kubenswrapper[4862]: E1128 11:43:50.598365 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b5266c9a26766fce2b92f95dff52d362a760f7baf1474cdcb33bd68570e096c0\\\"\"" pod="openstack/cinder-db-sync-5jcj5" podUID="b376a4e9-5456-4e8f-9c91-b222f38c30db" Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.623624 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.644780 4862 scope.go:117] "RemoveContainer" containerID="d16ba579580174081c4341491a21bdbf0d1f32acd583f99f26471ced2f0fe306" Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.666966 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-config\") pod \"82ea93c1-317d-4d0f-92a6-93872ef50d12\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.667031 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-combined-ca-bundle\") pod \"82ea93c1-317d-4d0f-92a6-93872ef50d12\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.667113 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlgwn\" (UniqueName: \"kubernetes.io/projected/82ea93c1-317d-4d0f-92a6-93872ef50d12-kube-api-access-dlgwn\") pod \"82ea93c1-317d-4d0f-92a6-93872ef50d12\" (UID: \"82ea93c1-317d-4d0f-92a6-93872ef50d12\") " Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.675751 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ea93c1-317d-4d0f-92a6-93872ef50d12-kube-api-access-dlgwn" (OuterVolumeSpecName: "kube-api-access-dlgwn") pod "82ea93c1-317d-4d0f-92a6-93872ef50d12" (UID: "82ea93c1-317d-4d0f-92a6-93872ef50d12"). InnerVolumeSpecName "kube-api-access-dlgwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:50 crc kubenswrapper[4862]: I1128 11:43:50.678553 4862 scope.go:117] "RemoveContainer" containerID="937b64afb7bb9f4152b8a58bafebe35f318d99aa762feb8feb4e8780be7a9b02" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.015316 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlgwn\" (UniqueName: \"kubernetes.io/projected/82ea93c1-317d-4d0f-92a6-93872ef50d12-kube-api-access-dlgwn\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.051577 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-config" (OuterVolumeSpecName: "config") pod "82ea93c1-317d-4d0f-92a6-93872ef50d12" (UID: "82ea93c1-317d-4d0f-92a6-93872ef50d12"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.052081 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82ea93c1-317d-4d0f-92a6-93872ef50d12" (UID: "82ea93c1-317d-4d0f-92a6-93872ef50d12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.081051 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-527fq"] Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.117271 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.117293 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ea93c1-317d-4d0f-92a6-93872ef50d12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.164512 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:43:51 crc kubenswrapper[4862]: W1128 11:43:51.165025 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod635693e3_57cb_479e_b4b0_947f81f10e97.slice/crio-fc934e164af1fa4fee2ac658b0e1f1625f2caed49251d893dba2e1d0b7d26ae7 WatchSource:0}: Error finding container fc934e164af1fa4fee2ac658b0e1f1625f2caed49251d893dba2e1d0b7d26ae7: Status 404 returned error can't find the container with id fc934e164af1fa4fee2ac658b0e1f1625f2caed49251d893dba2e1d0b7d26ae7 Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.573686 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-527fq" event={"ID":"57dc2822-b0b6-45d4-92de-5f601ebcf48a","Type":"ContainerStarted","Data":"0e41765288ac92ec73650d4eec94aa7bf446375875b6ddb400735a348f5b232a"} Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.574028 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-527fq" event={"ID":"57dc2822-b0b6-45d4-92de-5f601ebcf48a","Type":"ContainerStarted","Data":"eafab0ee159dbdc673a5d044683991b0dcc30a4ff916657e2c59906964b8add2"} Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.576628 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5lsz4" event={"ID":"eced1563-10aa-48f6-b79a-c017a3d4bba0","Type":"ContainerStarted","Data":"41c0bf7c013242ca225a1314624c3dc004d7184562ad1f7c5aa46b4bca21a2ae"} Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.578751 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerStarted","Data":"dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe"} Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.579582 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"635693e3-57cb-479e-b4b0-947f81f10e97","Type":"ContainerStarted","Data":"fc934e164af1fa4fee2ac658b0e1f1625f2caed49251d893dba2e1d0b7d26ae7"} Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.579625 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-47sh5" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.600188 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-527fq" podStartSLOduration=15.599417985 podStartE2EDuration="15.599417985s" podCreationTimestamp="2025-11-28 11:43:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:51.588452218 +0000 UTC m=+1309.120966139" watchObservedRunningTime="2025-11-28 11:43:51.599417985 +0000 UTC m=+1309.131931906" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.618016 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-5lsz4" podStartSLOduration=9.595017276 podStartE2EDuration="31.617996831s" podCreationTimestamp="2025-11-28 11:43:20 +0000 UTC" firstStartedPulling="2025-11-28 11:43:22.623017374 +0000 UTC m=+1280.155531295" lastFinishedPulling="2025-11-28 11:43:44.645996929 +0000 UTC m=+1302.178510850" observedRunningTime="2025-11-28 11:43:51.611577995 +0000 UTC m=+1309.144091906" watchObservedRunningTime="2025-11-28 11:43:51.617996831 +0000 UTC m=+1309.150510772" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.845495 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f8dc44d89-t4bpq"] Nov 28 11:43:51 crc kubenswrapper[4862]: E1128 11:43:51.846430 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerName="dnsmasq-dns" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.846450 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerName="dnsmasq-dns" Nov 28 11:43:51 crc kubenswrapper[4862]: E1128 11:43:51.846470 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerName="init" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.846476 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerName="init" Nov 28 11:43:51 crc kubenswrapper[4862]: E1128 11:43:51.846504 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ea93c1-317d-4d0f-92a6-93872ef50d12" containerName="neutron-db-sync" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.846511 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ea93c1-317d-4d0f-92a6-93872ef50d12" containerName="neutron-db-sync" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.846842 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ea93c1-317d-4d0f-92a6-93872ef50d12" containerName="neutron-db-sync" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.846871 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="c657ab4a-f20b-46a1-aa1b-35aed553d6c0" containerName="dnsmasq-dns" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.848198 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.932263 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f8dc44d89-t4bpq"] Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.966174 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6574c94c76-hq6jb"] Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.968311 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.970973 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.971260 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.971371 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.972954 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zbk6l" Nov 28 11:43:51 crc kubenswrapper[4862]: I1128 11:43:51.986512 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6574c94c76-hq6jb"] Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.043011 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-swift-storage-0\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.043132 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-svc\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.043169 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c5kv\" (UniqueName: \"kubernetes.io/projected/514236cd-f07b-41fd-8f58-f45c800642e6-kube-api-access-2c5kv\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.043233 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-nb\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.043315 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-sb\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.043393 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-config\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.129741 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:43:52 crc kubenswrapper[4862]: W1128 11:43:52.144610 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72c062a5_78c7_407a_8c0d_858a6fc301e2.slice/crio-24f3b7c97ea9cdda492c58ab7f2634be3b89d39f189e7f18b24ce84fe9eeb385 WatchSource:0}: Error finding container 24f3b7c97ea9cdda492c58ab7f2634be3b89d39f189e7f18b24ce84fe9eeb385: Status 404 returned error can't find the container with id 24f3b7c97ea9cdda492c58ab7f2634be3b89d39f189e7f18b24ce84fe9eeb385 Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145272 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-config\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145327 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-ovndb-tls-certs\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145346 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-combined-ca-bundle\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145408 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-swift-storage-0\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145441 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-svc\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145464 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c5kv\" (UniqueName: \"kubernetes.io/projected/514236cd-f07b-41fd-8f58-f45c800642e6-kube-api-access-2c5kv\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145485 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-nb\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145529 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-sb\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145551 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-httpd-config\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145580 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-config\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.145605 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsfsf\" (UniqueName: \"kubernetes.io/projected/225d091e-4cad-444e-aa3c-dcdf65dff654-kube-api-access-lsfsf\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.146619 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-swift-storage-0\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.146619 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-svc\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.147261 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-config\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.147414 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-sb\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.147434 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-nb\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.161544 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c5kv\" (UniqueName: \"kubernetes.io/projected/514236cd-f07b-41fd-8f58-f45c800642e6-kube-api-access-2c5kv\") pod \"dnsmasq-dns-f8dc44d89-t4bpq\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.247279 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-httpd-config\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.247358 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsfsf\" (UniqueName: \"kubernetes.io/projected/225d091e-4cad-444e-aa3c-dcdf65dff654-kube-api-access-lsfsf\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.247503 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-config\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.247540 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-ovndb-tls-certs\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.247585 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-combined-ca-bundle\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.251572 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-httpd-config\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.259309 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-config\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.259928 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-ovndb-tls-certs\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.260695 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-combined-ca-bundle\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.264024 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsfsf\" (UniqueName: \"kubernetes.io/projected/225d091e-4cad-444e-aa3c-dcdf65dff654-kube-api-access-lsfsf\") pod \"neutron-6574c94c76-hq6jb\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.317831 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.327539 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.647671 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"635693e3-57cb-479e-b4b0-947f81f10e97","Type":"ContainerStarted","Data":"14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68"} Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.651986 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"72c062a5-78c7-407a-8c0d-858a6fc301e2","Type":"ContainerStarted","Data":"24f3b7c97ea9cdda492c58ab7f2634be3b89d39f189e7f18b24ce84fe9eeb385"} Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.676050 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.676028424 podStartE2EDuration="8.676028424s" podCreationTimestamp="2025-11-28 11:43:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:52.663849612 +0000 UTC m=+1310.196363533" watchObservedRunningTime="2025-11-28 11:43:52.676028424 +0000 UTC m=+1310.208542345" Nov 28 11:43:52 crc kubenswrapper[4862]: I1128 11:43:52.948811 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f8dc44d89-t4bpq"] Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.108935 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6574c94c76-hq6jb"] Nov 28 11:43:53 crc kubenswrapper[4862]: W1128 11:43:53.110419 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod225d091e_4cad_444e_aa3c_dcdf65dff654.slice/crio-210790da95d404a97863f733039f054b289df93ea24cf3b4f3c90ae8a4b23403 WatchSource:0}: Error finding container 210790da95d404a97863f733039f054b289df93ea24cf3b4f3c90ae8a4b23403: Status 404 returned error can't find the container with id 210790da95d404a97863f733039f054b289df93ea24cf3b4f3c90ae8a4b23403 Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.664888 4862 generic.go:334] "Generic (PLEG): container finished" podID="eced1563-10aa-48f6-b79a-c017a3d4bba0" containerID="41c0bf7c013242ca225a1314624c3dc004d7184562ad1f7c5aa46b4bca21a2ae" exitCode=0 Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.665145 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5lsz4" event={"ID":"eced1563-10aa-48f6-b79a-c017a3d4bba0","Type":"ContainerDied","Data":"41c0bf7c013242ca225a1314624c3dc004d7184562ad1f7c5aa46b4bca21a2ae"} Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.668016 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerStarted","Data":"07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d"} Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.671579 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"635693e3-57cb-479e-b4b0-947f81f10e97","Type":"ContainerStarted","Data":"3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315"} Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.674223 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"72c062a5-78c7-407a-8c0d-858a6fc301e2","Type":"ContainerStarted","Data":"ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95"} Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.675257 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6574c94c76-hq6jb" event={"ID":"225d091e-4cad-444e-aa3c-dcdf65dff654","Type":"ContainerStarted","Data":"807efefdc1c9165b39db99614c63e8fc78929a6ac95c8615b88dc151742fe58e"} Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.675278 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6574c94c76-hq6jb" event={"ID":"225d091e-4cad-444e-aa3c-dcdf65dff654","Type":"ContainerStarted","Data":"210790da95d404a97863f733039f054b289df93ea24cf3b4f3c90ae8a4b23403"} Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.676207 4862 generic.go:334] "Generic (PLEG): container finished" podID="514236cd-f07b-41fd-8f58-f45c800642e6" containerID="f8f78fa6bc3eead90a890f3dd5d1f3fb7b5524a7998ad40d2954ed2bd92a945c" exitCode=0 Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.676291 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" event={"ID":"514236cd-f07b-41fd-8f58-f45c800642e6","Type":"ContainerDied","Data":"f8f78fa6bc3eead90a890f3dd5d1f3fb7b5524a7998ad40d2954ed2bd92a945c"} Nov 28 11:43:53 crc kubenswrapper[4862]: I1128 11:43:53.676354 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" event={"ID":"514236cd-f07b-41fd-8f58-f45c800642e6","Type":"ContainerStarted","Data":"9446f531d7af9edc09a967e9e43ec88821fcf9d8f8716a6984f8a3f5ea1a06ce"} Nov 28 11:43:54 crc kubenswrapper[4862]: E1128 11:43:54.586330 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57dc2822_b0b6_45d4_92de_5f601ebcf48a.slice/crio-0e41765288ac92ec73650d4eec94aa7bf446375875b6ddb400735a348f5b232a.scope\": RecentStats: unable to find data in memory cache]" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.695316 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"72c062a5-78c7-407a-8c0d-858a6fc301e2","Type":"ContainerStarted","Data":"b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba"} Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.701253 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6574c94c76-hq6jb" event={"ID":"225d091e-4cad-444e-aa3c-dcdf65dff654","Type":"ContainerStarted","Data":"c12815bdfd56949b92fac3d59058f4d61ac424748d6ad9871cb19fcac768ffac"} Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.701409 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.703001 4862 generic.go:334] "Generic (PLEG): container finished" podID="57dc2822-b0b6-45d4-92de-5f601ebcf48a" containerID="0e41765288ac92ec73650d4eec94aa7bf446375875b6ddb400735a348f5b232a" exitCode=0 Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.703045 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-527fq" event={"ID":"57dc2822-b0b6-45d4-92de-5f601ebcf48a","Type":"ContainerDied","Data":"0e41765288ac92ec73650d4eec94aa7bf446375875b6ddb400735a348f5b232a"} Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.705926 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" event={"ID":"514236cd-f07b-41fd-8f58-f45c800642e6","Type":"ContainerStarted","Data":"9b6db23f3e4130aa66413b92d4a49a64186b4cf841ad5c92cd550d02fbb0bb42"} Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.707303 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.719708 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=19.71969213 podStartE2EDuration="19.71969213s" podCreationTimestamp="2025-11-28 11:43:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:54.719558026 +0000 UTC m=+1312.252071937" watchObservedRunningTime="2025-11-28 11:43:54.71969213 +0000 UTC m=+1312.252206051" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.759661 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" podStartSLOduration=3.759643413 podStartE2EDuration="3.759643413s" podCreationTimestamp="2025-11-28 11:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:54.73840076 +0000 UTC m=+1312.270914681" watchObservedRunningTime="2025-11-28 11:43:54.759643413 +0000 UTC m=+1312.292157354" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.794292 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6dc889566f-77hs4"] Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.795820 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.802498 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.802558 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.811707 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6dc889566f-77hs4"] Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.814534 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6574c94c76-hq6jb" podStartSLOduration=3.814512656 podStartE2EDuration="3.814512656s" podCreationTimestamp="2025-11-28 11:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:43:54.785623082 +0000 UTC m=+1312.318137003" watchObservedRunningTime="2025-11-28 11:43:54.814512656 +0000 UTC m=+1312.347026567" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.896847 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.897201 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.906217 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-config\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.906272 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-internal-tls-certs\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.906424 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-public-tls-certs\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.906461 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-ovndb-tls-certs\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.906490 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-combined-ca-bundle\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.906539 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx27s\" (UniqueName: \"kubernetes.io/projected/a42b2a07-23e7-4281-b69b-e47635bcccf6-kube-api-access-jx27s\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.906709 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-httpd-config\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:54 crc kubenswrapper[4862]: I1128 11:43:54.974972 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.002392 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.008284 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-public-tls-certs\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.008403 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-ovndb-tls-certs\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.008498 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-combined-ca-bundle\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.008572 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx27s\" (UniqueName: \"kubernetes.io/projected/a42b2a07-23e7-4281-b69b-e47635bcccf6-kube-api-access-jx27s\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.008662 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-httpd-config\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.008739 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-config\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.008798 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-internal-tls-certs\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.015441 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-internal-tls-certs\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.015517 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-combined-ca-bundle\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.016039 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-public-tls-certs\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.016611 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-httpd-config\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.017146 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-config\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.019949 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-ovndb-tls-certs\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.024699 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx27s\" (UniqueName: \"kubernetes.io/projected/a42b2a07-23e7-4281-b69b-e47635bcccf6-kube-api-access-jx27s\") pod \"neutron-6dc889566f-77hs4\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.090410 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.129175 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.214792 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-config-data\") pod \"eced1563-10aa-48f6-b79a-c017a3d4bba0\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.214958 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-combined-ca-bundle\") pod \"eced1563-10aa-48f6-b79a-c017a3d4bba0\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.214999 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7qn8\" (UniqueName: \"kubernetes.io/projected/eced1563-10aa-48f6-b79a-c017a3d4bba0-kube-api-access-n7qn8\") pod \"eced1563-10aa-48f6-b79a-c017a3d4bba0\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.215038 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eced1563-10aa-48f6-b79a-c017a3d4bba0-logs\") pod \"eced1563-10aa-48f6-b79a-c017a3d4bba0\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.215121 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-scripts\") pod \"eced1563-10aa-48f6-b79a-c017a3d4bba0\" (UID: \"eced1563-10aa-48f6-b79a-c017a3d4bba0\") " Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.217350 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eced1563-10aa-48f6-b79a-c017a3d4bba0-logs" (OuterVolumeSpecName: "logs") pod "eced1563-10aa-48f6-b79a-c017a3d4bba0" (UID: "eced1563-10aa-48f6-b79a-c017a3d4bba0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.220046 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eced1563-10aa-48f6-b79a-c017a3d4bba0-kube-api-access-n7qn8" (OuterVolumeSpecName: "kube-api-access-n7qn8") pod "eced1563-10aa-48f6-b79a-c017a3d4bba0" (UID: "eced1563-10aa-48f6-b79a-c017a3d4bba0"). InnerVolumeSpecName "kube-api-access-n7qn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.232570 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-scripts" (OuterVolumeSpecName: "scripts") pod "eced1563-10aa-48f6-b79a-c017a3d4bba0" (UID: "eced1563-10aa-48f6-b79a-c017a3d4bba0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.288327 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eced1563-10aa-48f6-b79a-c017a3d4bba0" (UID: "eced1563-10aa-48f6-b79a-c017a3d4bba0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.302290 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-config-data" (OuterVolumeSpecName: "config-data") pod "eced1563-10aa-48f6-b79a-c017a3d4bba0" (UID: "eced1563-10aa-48f6-b79a-c017a3d4bba0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.317791 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.317818 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7qn8\" (UniqueName: \"kubernetes.io/projected/eced1563-10aa-48f6-b79a-c017a3d4bba0-kube-api-access-n7qn8\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.317831 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eced1563-10aa-48f6-b79a-c017a3d4bba0-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.317842 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.317850 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eced1563-10aa-48f6-b79a-c017a3d4bba0-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.695583 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6dc889566f-77hs4"] Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.725891 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5lsz4" event={"ID":"eced1563-10aa-48f6-b79a-c017a3d4bba0","Type":"ContainerDied","Data":"3baf996dbc637e47dd7542ac9960d71202c30cdbd3949613892fec7f6ae657f4"} Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.725990 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3baf996dbc637e47dd7542ac9960d71202c30cdbd3949613892fec7f6ae657f4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.726329 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5lsz4" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.726356 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.727047 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.845551 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-69477d9ff4-9mk7g"] Nov 28 11:43:55 crc kubenswrapper[4862]: E1128 11:43:55.845907 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eced1563-10aa-48f6-b79a-c017a3d4bba0" containerName="placement-db-sync" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.845921 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="eced1563-10aa-48f6-b79a-c017a3d4bba0" containerName="placement-db-sync" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.846148 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="eced1563-10aa-48f6-b79a-c017a3d4bba0" containerName="placement-db-sync" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.847936 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.854573 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.854924 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.855176 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.855467 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.859675 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-p6ckl" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.895749 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69477d9ff4-9mk7g"] Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.928244 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-scripts\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.928307 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-internal-tls-certs\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.929401 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-combined-ca-bundle\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.929457 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62671f-2a31-4980-93d9-2239defbfa57-logs\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.929632 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljnj4\" (UniqueName: \"kubernetes.io/projected/ca62671f-2a31-4980-93d9-2239defbfa57-kube-api-access-ljnj4\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.929674 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-public-tls-certs\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:55 crc kubenswrapper[4862]: I1128 11:43:55.929702 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-config-data\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.031779 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-combined-ca-bundle\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.031821 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62671f-2a31-4980-93d9-2239defbfa57-logs\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.031878 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljnj4\" (UniqueName: \"kubernetes.io/projected/ca62671f-2a31-4980-93d9-2239defbfa57-kube-api-access-ljnj4\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.031898 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-public-tls-certs\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.031916 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-config-data\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.031973 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-scripts\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.031999 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-internal-tls-certs\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.033443 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62671f-2a31-4980-93d9-2239defbfa57-logs\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.037547 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-public-tls-certs\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.038545 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-combined-ca-bundle\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.038880 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-scripts\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.039551 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-internal-tls-certs\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.039652 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-config-data\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.049323 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljnj4\" (UniqueName: \"kubernetes.io/projected/ca62671f-2a31-4980-93d9-2239defbfa57-kube-api-access-ljnj4\") pod \"placement-69477d9ff4-9mk7g\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.176714 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.370373 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.370745 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.425822 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.441660 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.735575 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 11:43:56 crc kubenswrapper[4862]: I1128 11:43:56.735605 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 11:43:58 crc kubenswrapper[4862]: I1128 11:43:58.856974 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:58 crc kubenswrapper[4862]: I1128 11:43:58.867176 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 11:43:59 crc kubenswrapper[4862]: W1128 11:43:59.318611 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda42b2a07_23e7_4281_b69b_e47635bcccf6.slice/crio-7445e87e625eae3d947ef68ff946f6761065fa7f935e089b7a3405bc6fcb97b1 WatchSource:0}: Error finding container 7445e87e625eae3d947ef68ff946f6761065fa7f935e089b7a3405bc6fcb97b1: Status 404 returned error can't find the container with id 7445e87e625eae3d947ef68ff946f6761065fa7f935e089b7a3405bc6fcb97b1 Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.608812 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.745878 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk7q4\" (UniqueName: \"kubernetes.io/projected/57dc2822-b0b6-45d4-92de-5f601ebcf48a-kube-api-access-pk7q4\") pod \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.745930 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-config-data\") pod \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.746046 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-credential-keys\") pod \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.746210 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-fernet-keys\") pod \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.746241 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-combined-ca-bundle\") pod \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.746265 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-scripts\") pod \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\" (UID: \"57dc2822-b0b6-45d4-92de-5f601ebcf48a\") " Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.751861 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "57dc2822-b0b6-45d4-92de-5f601ebcf48a" (UID: "57dc2822-b0b6-45d4-92de-5f601ebcf48a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.751914 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-scripts" (OuterVolumeSpecName: "scripts") pod "57dc2822-b0b6-45d4-92de-5f601ebcf48a" (UID: "57dc2822-b0b6-45d4-92de-5f601ebcf48a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.752730 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "57dc2822-b0b6-45d4-92de-5f601ebcf48a" (UID: "57dc2822-b0b6-45d4-92de-5f601ebcf48a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.752800 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57dc2822-b0b6-45d4-92de-5f601ebcf48a-kube-api-access-pk7q4" (OuterVolumeSpecName: "kube-api-access-pk7q4") pod "57dc2822-b0b6-45d4-92de-5f601ebcf48a" (UID: "57dc2822-b0b6-45d4-92de-5f601ebcf48a"). InnerVolumeSpecName "kube-api-access-pk7q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.764833 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-527fq" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.765972 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-527fq" event={"ID":"57dc2822-b0b6-45d4-92de-5f601ebcf48a","Type":"ContainerDied","Data":"eafab0ee159dbdc673a5d044683991b0dcc30a4ff916657e2c59906964b8add2"} Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.766008 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eafab0ee159dbdc673a5d044683991b0dcc30a4ff916657e2c59906964b8add2" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.766895 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc889566f-77hs4" event={"ID":"a42b2a07-23e7-4281-b69b-e47635bcccf6","Type":"ContainerStarted","Data":"9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab"} Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.766918 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc889566f-77hs4" event={"ID":"a42b2a07-23e7-4281-b69b-e47635bcccf6","Type":"ContainerStarted","Data":"7445e87e625eae3d947ef68ff946f6761065fa7f935e089b7a3405bc6fcb97b1"} Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.770899 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerStarted","Data":"46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce"} Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.776439 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-config-data" (OuterVolumeSpecName: "config-data") pod "57dc2822-b0b6-45d4-92de-5f601ebcf48a" (UID: "57dc2822-b0b6-45d4-92de-5f601ebcf48a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.779237 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.793700 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57dc2822-b0b6-45d4-92de-5f601ebcf48a" (UID: "57dc2822-b0b6-45d4-92de-5f601ebcf48a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.856729 4862 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.856879 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.856956 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.857012 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk7q4\" (UniqueName: \"kubernetes.io/projected/57dc2822-b0b6-45d4-92de-5f601ebcf48a-kube-api-access-pk7q4\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.857111 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.857191 4862 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57dc2822-b0b6-45d4-92de-5f601ebcf48a-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 11:43:59 crc kubenswrapper[4862]: W1128 11:43:59.886362 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca62671f_2a31_4980_93d9_2239defbfa57.slice/crio-d890a1ed5cddce80a6edf3b669ba45708f46953b7f4f422b9d8c1feb9ef58997 WatchSource:0}: Error finding container d890a1ed5cddce80a6edf3b669ba45708f46953b7f4f422b9d8c1feb9ef58997: Status 404 returned error can't find the container with id d890a1ed5cddce80a6edf3b669ba45708f46953b7f4f422b9d8c1feb9ef58997 Nov 28 11:43:59 crc kubenswrapper[4862]: I1128 11:43:59.887197 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69477d9ff4-9mk7g"] Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.715249 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7655db979f-pdkkm"] Nov 28 11:44:00 crc kubenswrapper[4862]: E1128 11:44:00.715985 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dc2822-b0b6-45d4-92de-5f601ebcf48a" containerName="keystone-bootstrap" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.716004 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dc2822-b0b6-45d4-92de-5f601ebcf48a" containerName="keystone-bootstrap" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.716187 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="57dc2822-b0b6-45d4-92de-5f601ebcf48a" containerName="keystone-bootstrap" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.716799 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.719270 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gn5r2" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.724428 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.724761 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.724885 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.725037 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.725615 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.733846 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7655db979f-pdkkm"] Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.774399 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-fernet-keys\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.774458 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-config-data\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.774481 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-public-tls-certs\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.774524 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b9k7\" (UniqueName: \"kubernetes.io/projected/0c30075d-2cd8-4f5a-9168-33c55459f636-kube-api-access-9b9k7\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.774558 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-scripts\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.774576 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-combined-ca-bundle\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.774594 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-internal-tls-certs\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.774614 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-credential-keys\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.783968 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69477d9ff4-9mk7g" event={"ID":"ca62671f-2a31-4980-93d9-2239defbfa57","Type":"ContainerStarted","Data":"a7c34e84cffc71a4ccf07325940e732e6d5ffc06d972e6e93c18df3db794e0b9"} Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.784009 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69477d9ff4-9mk7g" event={"ID":"ca62671f-2a31-4980-93d9-2239defbfa57","Type":"ContainerStarted","Data":"ec72be9e8e9e4302a004ec5624c9e7d3c1b7137a5f5cc1b75511dcf0b235953b"} Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.784027 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69477d9ff4-9mk7g" event={"ID":"ca62671f-2a31-4980-93d9-2239defbfa57","Type":"ContainerStarted","Data":"d890a1ed5cddce80a6edf3b669ba45708f46953b7f4f422b9d8c1feb9ef58997"} Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.784330 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.784368 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.788458 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dnh4p" event={"ID":"4be2cd5e-1345-4212-9563-449a665090f1","Type":"ContainerStarted","Data":"7e861fa3db125e3ec49410b4062ef8c9f7b09f017f185673a8589264da58f5ce"} Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.790924 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc889566f-77hs4" event={"ID":"a42b2a07-23e7-4281-b69b-e47635bcccf6","Type":"ContainerStarted","Data":"76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f"} Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.791633 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.807597 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-69477d9ff4-9mk7g" podStartSLOduration=5.807578417 podStartE2EDuration="5.807578417s" podCreationTimestamp="2025-11-28 11:43:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:00.801853432 +0000 UTC m=+1318.334367353" watchObservedRunningTime="2025-11-28 11:44:00.807578417 +0000 UTC m=+1318.340092338" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.832172 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6dc889566f-77hs4" podStartSLOduration=6.832153676 podStartE2EDuration="6.832153676s" podCreationTimestamp="2025-11-28 11:43:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:00.822468157 +0000 UTC m=+1318.354982078" watchObservedRunningTime="2025-11-28 11:44:00.832153676 +0000 UTC m=+1318.364667597" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.849082 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dnh4p" podStartSLOduration=2.84813873 podStartE2EDuration="40.849062884s" podCreationTimestamp="2025-11-28 11:43:20 +0000 UTC" firstStartedPulling="2025-11-28 11:43:22.326554823 +0000 UTC m=+1279.859068744" lastFinishedPulling="2025-11-28 11:44:00.327478977 +0000 UTC m=+1317.859992898" observedRunningTime="2025-11-28 11:44:00.844102301 +0000 UTC m=+1318.376616222" watchObservedRunningTime="2025-11-28 11:44:00.849062884 +0000 UTC m=+1318.381576805" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.876577 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-internal-tls-certs\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.876633 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-credential-keys\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.876761 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-fernet-keys\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.876830 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-config-data\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.876851 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-public-tls-certs\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.876915 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b9k7\" (UniqueName: \"kubernetes.io/projected/0c30075d-2cd8-4f5a-9168-33c55459f636-kube-api-access-9b9k7\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.876971 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-scripts\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.876994 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-combined-ca-bundle\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.884929 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-config-data\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.885549 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-scripts\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.886024 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-fernet-keys\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.886032 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-public-tls-certs\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.886253 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-internal-tls-certs\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.886804 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-credential-keys\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.888546 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-combined-ca-bundle\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:00 crc kubenswrapper[4862]: I1128 11:44:00.900667 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b9k7\" (UniqueName: \"kubernetes.io/projected/0c30075d-2cd8-4f5a-9168-33c55459f636-kube-api-access-9b9k7\") pod \"keystone-7655db979f-pdkkm\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:01 crc kubenswrapper[4862]: I1128 11:44:01.041593 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:01 crc kubenswrapper[4862]: I1128 11:44:01.540851 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7655db979f-pdkkm"] Nov 28 11:44:01 crc kubenswrapper[4862]: I1128 11:44:01.820316 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7655db979f-pdkkm" event={"ID":"0c30075d-2cd8-4f5a-9168-33c55459f636","Type":"ContainerStarted","Data":"b508df719851845b47a039b4e7bfe0523253f825c7e85f7002bc75e3ce377fbc"} Nov 28 11:44:02 crc kubenswrapper[4862]: I1128 11:44:02.036543 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 11:44:02 crc kubenswrapper[4862]: I1128 11:44:02.320295 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:44:02 crc kubenswrapper[4862]: I1128 11:44:02.394392 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f5d458b55-bfmjn"] Nov 28 11:44:02 crc kubenswrapper[4862]: I1128 11:44:02.394624 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" podUID="0bca720b-962c-48e5-9861-406c78a5b202" containerName="dnsmasq-dns" containerID="cri-o://b570c3eb66bb4f0f229dea23e06f021c255107f84372e2683901f74f1fa171e4" gracePeriod=10 Nov 28 11:44:02 crc kubenswrapper[4862]: I1128 11:44:02.833271 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7655db979f-pdkkm" event={"ID":"0c30075d-2cd8-4f5a-9168-33c55459f636","Type":"ContainerStarted","Data":"5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b"} Nov 28 11:44:02 crc kubenswrapper[4862]: I1128 11:44:02.833663 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:02 crc kubenswrapper[4862]: I1128 11:44:02.837687 4862 generic.go:334] "Generic (PLEG): container finished" podID="0bca720b-962c-48e5-9861-406c78a5b202" containerID="b570c3eb66bb4f0f229dea23e06f021c255107f84372e2683901f74f1fa171e4" exitCode=0 Nov 28 11:44:02 crc kubenswrapper[4862]: I1128 11:44:02.854366 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7655db979f-pdkkm" podStartSLOduration=2.854347555 podStartE2EDuration="2.854347555s" podCreationTimestamp="2025-11-28 11:44:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:02.853810949 +0000 UTC m=+1320.386324890" watchObservedRunningTime="2025-11-28 11:44:02.854347555 +0000 UTC m=+1320.386861476" Nov 28 11:44:02 crc kubenswrapper[4862]: I1128 11:44:02.867519 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" event={"ID":"0bca720b-962c-48e5-9861-406c78a5b202","Type":"ContainerDied","Data":"b570c3eb66bb4f0f229dea23e06f021c255107f84372e2683901f74f1fa171e4"} Nov 28 11:44:03 crc kubenswrapper[4862]: I1128 11:44:03.960591 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.062835 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-config\") pod \"0bca720b-962c-48e5-9861-406c78a5b202\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.062915 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-swift-storage-0\") pod \"0bca720b-962c-48e5-9861-406c78a5b202\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.063027 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-svc\") pod \"0bca720b-962c-48e5-9861-406c78a5b202\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.063219 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-nb\") pod \"0bca720b-962c-48e5-9861-406c78a5b202\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.063250 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-sb\") pod \"0bca720b-962c-48e5-9861-406c78a5b202\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.063296 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr56q\" (UniqueName: \"kubernetes.io/projected/0bca720b-962c-48e5-9861-406c78a5b202-kube-api-access-rr56q\") pod \"0bca720b-962c-48e5-9861-406c78a5b202\" (UID: \"0bca720b-962c-48e5-9861-406c78a5b202\") " Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.089294 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bca720b-962c-48e5-9861-406c78a5b202-kube-api-access-rr56q" (OuterVolumeSpecName: "kube-api-access-rr56q") pod "0bca720b-962c-48e5-9861-406c78a5b202" (UID: "0bca720b-962c-48e5-9861-406c78a5b202"). InnerVolumeSpecName "kube-api-access-rr56q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.117900 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0bca720b-962c-48e5-9861-406c78a5b202" (UID: "0bca720b-962c-48e5-9861-406c78a5b202"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.139751 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-config" (OuterVolumeSpecName: "config") pod "0bca720b-962c-48e5-9861-406c78a5b202" (UID: "0bca720b-962c-48e5-9861-406c78a5b202"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.141953 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0bca720b-962c-48e5-9861-406c78a5b202" (UID: "0bca720b-962c-48e5-9861-406c78a5b202"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.147507 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0bca720b-962c-48e5-9861-406c78a5b202" (UID: "0bca720b-962c-48e5-9861-406c78a5b202"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.148514 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0bca720b-962c-48e5-9861-406c78a5b202" (UID: "0bca720b-962c-48e5-9861-406c78a5b202"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.165887 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.165923 4862 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.165938 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.165946 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.165955 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bca720b-962c-48e5-9861-406c78a5b202-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.165964 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr56q\" (UniqueName: \"kubernetes.io/projected/0bca720b-962c-48e5-9861-406c78a5b202-kube-api-access-rr56q\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.868813 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" event={"ID":"0bca720b-962c-48e5-9861-406c78a5b202","Type":"ContainerDied","Data":"dc4975e4749b7934bc65feb740a704378b7be1920f999976219d27caaa9eae16"} Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.868845 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5d458b55-bfmjn" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.868885 4862 scope.go:117] "RemoveContainer" containerID="b570c3eb66bb4f0f229dea23e06f021c255107f84372e2683901f74f1fa171e4" Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.874653 4862 generic.go:334] "Generic (PLEG): container finished" podID="4be2cd5e-1345-4212-9563-449a665090f1" containerID="7e861fa3db125e3ec49410b4062ef8c9f7b09f017f185673a8589264da58f5ce" exitCode=0 Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.874706 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dnh4p" event={"ID":"4be2cd5e-1345-4212-9563-449a665090f1","Type":"ContainerDied","Data":"7e861fa3db125e3ec49410b4062ef8c9f7b09f017f185673a8589264da58f5ce"} Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.877876 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5jcj5" event={"ID":"b376a4e9-5456-4e8f-9c91-b222f38c30db","Type":"ContainerStarted","Data":"24e2b7bdd8bb596c772aeefbf5a89bcacb1c66292c794bd6e2daecfa357f1740"} Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.916498 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f5d458b55-bfmjn"] Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.927757 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f5d458b55-bfmjn"] Nov 28 11:44:04 crc kubenswrapper[4862]: I1128 11:44:04.936548 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-5jcj5" podStartSLOduration=3.652014672 podStartE2EDuration="44.936527972s" podCreationTimestamp="2025-11-28 11:43:20 +0000 UTC" firstStartedPulling="2025-11-28 11:43:22.521632142 +0000 UTC m=+1280.054146063" lastFinishedPulling="2025-11-28 11:44:03.806145442 +0000 UTC m=+1321.338659363" observedRunningTime="2025-11-28 11:44:04.924154435 +0000 UTC m=+1322.456668356" watchObservedRunningTime="2025-11-28 11:44:04.936527972 +0000 UTC m=+1322.469041913" Nov 28 11:44:06 crc kubenswrapper[4862]: I1128 11:44:06.852378 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bca720b-962c-48e5-9861-406c78a5b202" path="/var/lib/kubelet/pods/0bca720b-962c-48e5-9861-406c78a5b202/volumes" Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.078894 4862 scope.go:117] "RemoveContainer" containerID="18e893787b962c2c301c113217c240503caa6893de81f9152dd44079187812f3" Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.875812 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.917652 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dnh4p" event={"ID":"4be2cd5e-1345-4212-9563-449a665090f1","Type":"ContainerDied","Data":"1e18d482ffa1b6d12eaece1875d3baa5e50a3d241ba5e9fe08357138da42398f"} Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.917695 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e18d482ffa1b6d12eaece1875d3baa5e50a3d241ba5e9fe08357138da42398f" Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.917725 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dnh4p" Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.937644 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-combined-ca-bundle\") pod \"4be2cd5e-1345-4212-9563-449a665090f1\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.937804 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-db-sync-config-data\") pod \"4be2cd5e-1345-4212-9563-449a665090f1\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.937864 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6pz5\" (UniqueName: \"kubernetes.io/projected/4be2cd5e-1345-4212-9563-449a665090f1-kube-api-access-g6pz5\") pod \"4be2cd5e-1345-4212-9563-449a665090f1\" (UID: \"4be2cd5e-1345-4212-9563-449a665090f1\") " Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.942745 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4be2cd5e-1345-4212-9563-449a665090f1" (UID: "4be2cd5e-1345-4212-9563-449a665090f1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.952749 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4be2cd5e-1345-4212-9563-449a665090f1-kube-api-access-g6pz5" (OuterVolumeSpecName: "kube-api-access-g6pz5") pod "4be2cd5e-1345-4212-9563-449a665090f1" (UID: "4be2cd5e-1345-4212-9563-449a665090f1"). InnerVolumeSpecName "kube-api-access-g6pz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:07 crc kubenswrapper[4862]: I1128 11:44:07.983775 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4be2cd5e-1345-4212-9563-449a665090f1" (UID: "4be2cd5e-1345-4212-9563-449a665090f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.041974 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.042050 4862 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4be2cd5e-1345-4212-9563-449a665090f1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.042112 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6pz5\" (UniqueName: \"kubernetes.io/projected/4be2cd5e-1345-4212-9563-449a665090f1-kube-api-access-g6pz5\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.937916 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerStarted","Data":"97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898"} Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.938450 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="ceilometer-central-agent" containerID="cri-o://dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe" gracePeriod=30 Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.938736 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.939052 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="proxy-httpd" containerID="cri-o://97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898" gracePeriod=30 Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.939145 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="sg-core" containerID="cri-o://46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce" gracePeriod=30 Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.939207 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="ceilometer-notification-agent" containerID="cri-o://07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d" gracePeriod=30 Nov 28 11:44:08 crc kubenswrapper[4862]: I1128 11:44:08.979388 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.287667742 podStartE2EDuration="48.979360543s" podCreationTimestamp="2025-11-28 11:43:20 +0000 UTC" firstStartedPulling="2025-11-28 11:43:22.690998609 +0000 UTC m=+1280.223512520" lastFinishedPulling="2025-11-28 11:44:08.38269139 +0000 UTC m=+1325.915205321" observedRunningTime="2025-11-28 11:44:08.970055694 +0000 UTC m=+1326.502569645" watchObservedRunningTime="2025-11-28 11:44:08.979360543 +0000 UTC m=+1326.511874484" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.259483 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6df5c754df-6lvmr"] Nov 28 11:44:09 crc kubenswrapper[4862]: E1128 11:44:09.259907 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bca720b-962c-48e5-9861-406c78a5b202" containerName="init" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.259922 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bca720b-962c-48e5-9861-406c78a5b202" containerName="init" Nov 28 11:44:09 crc kubenswrapper[4862]: E1128 11:44:09.259937 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bca720b-962c-48e5-9861-406c78a5b202" containerName="dnsmasq-dns" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.259945 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bca720b-962c-48e5-9861-406c78a5b202" containerName="dnsmasq-dns" Nov 28 11:44:09 crc kubenswrapper[4862]: E1128 11:44:09.259964 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4be2cd5e-1345-4212-9563-449a665090f1" containerName="barbican-db-sync" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.259974 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4be2cd5e-1345-4212-9563-449a665090f1" containerName="barbican-db-sync" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.260217 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="4be2cd5e-1345-4212-9563-449a665090f1" containerName="barbican-db-sync" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.260249 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bca720b-962c-48e5-9861-406c78a5b202" containerName="dnsmasq-dns" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.261570 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.277862 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-combined-ca-bundle\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.277995 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.278040 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data-custom\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.278343 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5203b21-897c-4331-a878-1d5c949a3e75-logs\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.278383 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkdk8\" (UniqueName: \"kubernetes.io/projected/d5203b21-897c-4331-a878-1d5c949a3e75-kube-api-access-dkdk8\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.301471 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.301603 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.301758 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-nvwvc" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.315029 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6df5c754df-6lvmr"] Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.380076 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-combined-ca-bundle\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.380509 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.380542 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data-custom\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.380576 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5203b21-897c-4331-a878-1d5c949a3e75-logs\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.380596 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkdk8\" (UniqueName: \"kubernetes.io/projected/d5203b21-897c-4331-a878-1d5c949a3e75-kube-api-access-dkdk8\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.381279 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5203b21-897c-4331-a878-1d5c949a3e75-logs\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.398878 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-combined-ca-bundle\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.399779 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data-custom\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.407430 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.409032 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7756445dd6-vwrsv"] Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.422311 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7756445dd6-vwrsv"] Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.431942 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkdk8\" (UniqueName: \"kubernetes.io/projected/d5203b21-897c-4331-a878-1d5c949a3e75-kube-api-access-dkdk8\") pod \"barbican-worker-6df5c754df-6lvmr\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.456405 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.461402 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.483752 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data-custom\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.483861 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-combined-ca-bundle\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.483943 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb5sl\" (UniqueName: \"kubernetes.io/projected/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-kube-api-access-hb5sl\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.484015 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-logs\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.484053 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.561421 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf96b7dc5-h7x89"] Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.563250 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587234 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587286 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-sb\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587319 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-svc\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587352 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data-custom\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587400 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-combined-ca-bundle\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587445 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-config\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587467 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9dmc\" (UniqueName: \"kubernetes.io/projected/701596e6-38ae-4964-839d-44118647085e-kube-api-access-t9dmc\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587489 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb5sl\" (UniqueName: \"kubernetes.io/projected/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-kube-api-access-hb5sl\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587508 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-swift-storage-0\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587535 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-nb\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587565 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-logs\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.587949 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-logs\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.592004 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.596986 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-combined-ca-bundle\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.616639 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.616748 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data-custom\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.639348 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb5sl\" (UniqueName: \"kubernetes.io/projected/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-kube-api-access-hb5sl\") pod \"barbican-keystone-listener-7756445dd6-vwrsv\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.666118 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf96b7dc5-h7x89"] Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.687121 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-599f8c654b-clr4b"] Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.688480 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.688564 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9dmc\" (UniqueName: \"kubernetes.io/projected/701596e6-38ae-4964-839d-44118647085e-kube-api-access-t9dmc\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.688599 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-swift-storage-0\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.688627 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-nb\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.688666 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-sb\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.688694 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-svc\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.688774 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-config\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.689621 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-config\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.690151 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-nb\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.690564 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-svc\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.690735 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-swift-storage-0\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.690889 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-sb\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.695314 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.703489 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-599f8c654b-clr4b"] Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.706389 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9dmc\" (UniqueName: \"kubernetes.io/projected/701596e6-38ae-4964-839d-44118647085e-kube-api-access-t9dmc\") pod \"dnsmasq-dns-cf96b7dc5-h7x89\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.790998 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-combined-ca-bundle\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.792270 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.792877 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-logs\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.792905 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7pnf\" (UniqueName: \"kubernetes.io/projected/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-kube-api-access-v7pnf\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.793042 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data-custom\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.834426 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.892718 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.894388 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7pnf\" (UniqueName: \"kubernetes.io/projected/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-kube-api-access-v7pnf\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.894444 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data-custom\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.894550 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-combined-ca-bundle\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.894607 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.894659 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-logs\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.895987 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-logs\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.905157 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-combined-ca-bundle\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.912063 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data-custom\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.912284 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.914347 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7pnf\" (UniqueName: \"kubernetes.io/projected/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-kube-api-access-v7pnf\") pod \"barbican-api-599f8c654b-clr4b\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.989548 4862 generic.go:334] "Generic (PLEG): container finished" podID="5c005a75-9925-4968-bb0f-40c16c805d35" containerID="97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898" exitCode=0 Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.989572 4862 generic.go:334] "Generic (PLEG): container finished" podID="5c005a75-9925-4968-bb0f-40c16c805d35" containerID="46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce" exitCode=2 Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.989582 4862 generic.go:334] "Generic (PLEG): container finished" podID="5c005a75-9925-4968-bb0f-40c16c805d35" containerID="dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe" exitCode=0 Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.989634 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerDied","Data":"97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898"} Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.989659 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerDied","Data":"46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce"} Nov 28 11:44:09 crc kubenswrapper[4862]: I1128 11:44:09.989671 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerDied","Data":"dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe"} Nov 28 11:44:10 crc kubenswrapper[4862]: I1128 11:44:10.026240 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:10 crc kubenswrapper[4862]: I1128 11:44:10.118958 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6df5c754df-6lvmr"] Nov 28 11:44:10 crc kubenswrapper[4862]: I1128 11:44:10.362463 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7756445dd6-vwrsv"] Nov 28 11:44:10 crc kubenswrapper[4862]: W1128 11:44:10.370263 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4f23c53_e6f8_4aca_a1e4_969a58b70a38.slice/crio-048866d2590e7fca7d149171a3413cf0ef36cd649289e039311ed45d486fc1fa WatchSource:0}: Error finding container 048866d2590e7fca7d149171a3413cf0ef36cd649289e039311ed45d486fc1fa: Status 404 returned error can't find the container with id 048866d2590e7fca7d149171a3413cf0ef36cd649289e039311ed45d486fc1fa Nov 28 11:44:10 crc kubenswrapper[4862]: I1128 11:44:10.454613 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf96b7dc5-h7x89"] Nov 28 11:44:10 crc kubenswrapper[4862]: W1128 11:44:10.460832 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod701596e6_38ae_4964_839d_44118647085e.slice/crio-8a696d0b6b75425a23786eed252c30d3ed5a29a92a1dfe64dd05657844f016bc WatchSource:0}: Error finding container 8a696d0b6b75425a23786eed252c30d3ed5a29a92a1dfe64dd05657844f016bc: Status 404 returned error can't find the container with id 8a696d0b6b75425a23786eed252c30d3ed5a29a92a1dfe64dd05657844f016bc Nov 28 11:44:10 crc kubenswrapper[4862]: I1128 11:44:10.541410 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-599f8c654b-clr4b"] Nov 28 11:44:10 crc kubenswrapper[4862]: W1128 11:44:10.557594 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0f90517_7935_4112_bcc9_07ac2e3cd2f8.slice/crio-33e7682ec7f5c765a4e00f3fa3d4a98d1c6bd42fc1113a9e2c1bd7b76b348277 WatchSource:0}: Error finding container 33e7682ec7f5c765a4e00f3fa3d4a98d1c6bd42fc1113a9e2c1bd7b76b348277: Status 404 returned error can't find the container with id 33e7682ec7f5c765a4e00f3fa3d4a98d1c6bd42fc1113a9e2c1bd7b76b348277 Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.005414 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" event={"ID":"e4f23c53-e6f8-4aca-a1e4-969a58b70a38","Type":"ContainerStarted","Data":"048866d2590e7fca7d149171a3413cf0ef36cd649289e039311ed45d486fc1fa"} Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.008024 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-599f8c654b-clr4b" event={"ID":"e0f90517-7935-4112-bcc9-07ac2e3cd2f8","Type":"ContainerStarted","Data":"6b20364c8c922514b37f40471f7578b3c0622702bd785eb84a5903eb3f50eb68"} Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.008061 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-599f8c654b-clr4b" event={"ID":"e0f90517-7935-4112-bcc9-07ac2e3cd2f8","Type":"ContainerStarted","Data":"33e7682ec7f5c765a4e00f3fa3d4a98d1c6bd42fc1113a9e2c1bd7b76b348277"} Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.009598 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6df5c754df-6lvmr" event={"ID":"d5203b21-897c-4331-a878-1d5c949a3e75","Type":"ContainerStarted","Data":"749f84f1d79e4dc9a76a2699b9668ed8f2dee3349b07cfde631e2b369b80b82f"} Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.011756 4862 generic.go:334] "Generic (PLEG): container finished" podID="b376a4e9-5456-4e8f-9c91-b222f38c30db" containerID="24e2b7bdd8bb596c772aeefbf5a89bcacb1c66292c794bd6e2daecfa357f1740" exitCode=0 Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.011814 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5jcj5" event={"ID":"b376a4e9-5456-4e8f-9c91-b222f38c30db","Type":"ContainerDied","Data":"24e2b7bdd8bb596c772aeefbf5a89bcacb1c66292c794bd6e2daecfa357f1740"} Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.014335 4862 generic.go:334] "Generic (PLEG): container finished" podID="701596e6-38ae-4964-839d-44118647085e" containerID="d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c" exitCode=0 Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.014368 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" event={"ID":"701596e6-38ae-4964-839d-44118647085e","Type":"ContainerDied","Data":"d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c"} Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.014387 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" event={"ID":"701596e6-38ae-4964-839d-44118647085e","Type":"ContainerStarted","Data":"8a696d0b6b75425a23786eed252c30d3ed5a29a92a1dfe64dd05657844f016bc"} Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.932872 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-857784bb5b-4pc8q"] Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.934764 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.938859 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.939183 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 28 11:44:11 crc kubenswrapper[4862]: I1128 11:44:11.945962 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-857784bb5b-4pc8q"] Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.023802 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" event={"ID":"701596e6-38ae-4964-839d-44118647085e","Type":"ContainerStarted","Data":"f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8"} Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.025731 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-599f8c654b-clr4b" event={"ID":"e0f90517-7935-4112-bcc9-07ac2e3cd2f8","Type":"ContainerStarted","Data":"adb727d122542042ce3d39254042b51e58e9fb18c8bdcf839eb1e8cf626fd5cd"} Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.036446 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data-custom\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.036787 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-internal-tls-certs\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.036872 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91afecb8-a07b-43e5-92e4-4e86cce7bcca-logs\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.036905 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9hrc\" (UniqueName: \"kubernetes.io/projected/91afecb8-a07b-43e5-92e4-4e86cce7bcca-kube-api-access-c9hrc\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.036965 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-combined-ca-bundle\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.037291 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-public-tls-certs\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.037406 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.044640 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" podStartSLOduration=3.044621861 podStartE2EDuration="3.044621861s" podCreationTimestamp="2025-11-28 11:44:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:12.040583445 +0000 UTC m=+1329.573097376" watchObservedRunningTime="2025-11-28 11:44:12.044621861 +0000 UTC m=+1329.577135782" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.070119 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-599f8c654b-clr4b" podStartSLOduration=3.070079856 podStartE2EDuration="3.070079856s" podCreationTimestamp="2025-11-28 11:44:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:12.060416267 +0000 UTC m=+1329.592930188" watchObservedRunningTime="2025-11-28 11:44:12.070079856 +0000 UTC m=+1329.602593777" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.142285 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-public-tls-certs\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.143014 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.143061 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data-custom\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.143139 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-internal-tls-certs\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.143172 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91afecb8-a07b-43e5-92e4-4e86cce7bcca-logs\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.143197 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9hrc\" (UniqueName: \"kubernetes.io/projected/91afecb8-a07b-43e5-92e4-4e86cce7bcca-kube-api-access-c9hrc\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.143222 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-combined-ca-bundle\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.143637 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91afecb8-a07b-43e5-92e4-4e86cce7bcca-logs\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.150951 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-combined-ca-bundle\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.151664 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-public-tls-certs\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.159575 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-internal-tls-certs\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.161197 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.168696 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data-custom\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.175718 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9hrc\" (UniqueName: \"kubernetes.io/projected/91afecb8-a07b-43e5-92e4-4e86cce7bcca-kube-api-access-c9hrc\") pod \"barbican-api-857784bb5b-4pc8q\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.349296 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.408314 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.448122 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xdmc\" (UniqueName: \"kubernetes.io/projected/b376a4e9-5456-4e8f-9c91-b222f38c30db-kube-api-access-5xdmc\") pod \"b376a4e9-5456-4e8f-9c91-b222f38c30db\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.448184 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b376a4e9-5456-4e8f-9c91-b222f38c30db-etc-machine-id\") pod \"b376a4e9-5456-4e8f-9c91-b222f38c30db\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.448230 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-scripts\") pod \"b376a4e9-5456-4e8f-9c91-b222f38c30db\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.448287 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-config-data\") pod \"b376a4e9-5456-4e8f-9c91-b222f38c30db\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.448361 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-combined-ca-bundle\") pod \"b376a4e9-5456-4e8f-9c91-b222f38c30db\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.448394 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-db-sync-config-data\") pod \"b376a4e9-5456-4e8f-9c91-b222f38c30db\" (UID: \"b376a4e9-5456-4e8f-9c91-b222f38c30db\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.449055 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b376a4e9-5456-4e8f-9c91-b222f38c30db-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b376a4e9-5456-4e8f-9c91-b222f38c30db" (UID: "b376a4e9-5456-4e8f-9c91-b222f38c30db"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.460039 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-scripts" (OuterVolumeSpecName: "scripts") pod "b376a4e9-5456-4e8f-9c91-b222f38c30db" (UID: "b376a4e9-5456-4e8f-9c91-b222f38c30db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.468371 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b376a4e9-5456-4e8f-9c91-b222f38c30db" (UID: "b376a4e9-5456-4e8f-9c91-b222f38c30db"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.472468 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b376a4e9-5456-4e8f-9c91-b222f38c30db-kube-api-access-5xdmc" (OuterVolumeSpecName: "kube-api-access-5xdmc") pod "b376a4e9-5456-4e8f-9c91-b222f38c30db" (UID: "b376a4e9-5456-4e8f-9c91-b222f38c30db"). InnerVolumeSpecName "kube-api-access-5xdmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.531395 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b376a4e9-5456-4e8f-9c91-b222f38c30db" (UID: "b376a4e9-5456-4e8f-9c91-b222f38c30db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.547308 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-config-data" (OuterVolumeSpecName: "config-data") pod "b376a4e9-5456-4e8f-9c91-b222f38c30db" (UID: "b376a4e9-5456-4e8f-9c91-b222f38c30db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.556478 4862 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b376a4e9-5456-4e8f-9c91-b222f38c30db-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.556506 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.556514 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.556524 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.556534 4862 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b376a4e9-5456-4e8f-9c91-b222f38c30db-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.556544 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xdmc\" (UniqueName: \"kubernetes.io/projected/b376a4e9-5456-4e8f-9c91-b222f38c30db-kube-api-access-5xdmc\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.770431 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.860309 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-log-httpd\") pod \"5c005a75-9925-4968-bb0f-40c16c805d35\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.860373 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-sg-core-conf-yaml\") pod \"5c005a75-9925-4968-bb0f-40c16c805d35\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.860423 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-config-data\") pod \"5c005a75-9925-4968-bb0f-40c16c805d35\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.860478 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pskx\" (UniqueName: \"kubernetes.io/projected/5c005a75-9925-4968-bb0f-40c16c805d35-kube-api-access-2pskx\") pod \"5c005a75-9925-4968-bb0f-40c16c805d35\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.860502 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-combined-ca-bundle\") pod \"5c005a75-9925-4968-bb0f-40c16c805d35\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.860576 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-scripts\") pod \"5c005a75-9925-4968-bb0f-40c16c805d35\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.860745 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-run-httpd\") pod \"5c005a75-9925-4968-bb0f-40c16c805d35\" (UID: \"5c005a75-9925-4968-bb0f-40c16c805d35\") " Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.861792 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5c005a75-9925-4968-bb0f-40c16c805d35" (UID: "5c005a75-9925-4968-bb0f-40c16c805d35"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.862980 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5c005a75-9925-4968-bb0f-40c16c805d35" (UID: "5c005a75-9925-4968-bb0f-40c16c805d35"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.865682 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c005a75-9925-4968-bb0f-40c16c805d35-kube-api-access-2pskx" (OuterVolumeSpecName: "kube-api-access-2pskx") pod "5c005a75-9925-4968-bb0f-40c16c805d35" (UID: "5c005a75-9925-4968-bb0f-40c16c805d35"). InnerVolumeSpecName "kube-api-access-2pskx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.870813 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-scripts" (OuterVolumeSpecName: "scripts") pod "5c005a75-9925-4968-bb0f-40c16c805d35" (UID: "5c005a75-9925-4968-bb0f-40c16c805d35"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.911124 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5c005a75-9925-4968-bb0f-40c16c805d35" (UID: "5c005a75-9925-4968-bb0f-40c16c805d35"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.939680 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-857784bb5b-4pc8q"] Nov 28 11:44:12 crc kubenswrapper[4862]: W1128 11:44:12.955875 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91afecb8_a07b_43e5_92e4_4e86cce7bcca.slice/crio-807c30986f85c37de51a7b74d0c8e012c3c5178a714e192cb5cce33e8d23e0f1 WatchSource:0}: Error finding container 807c30986f85c37de51a7b74d0c8e012c3c5178a714e192cb5cce33e8d23e0f1: Status 404 returned error can't find the container with id 807c30986f85c37de51a7b74d0c8e012c3c5178a714e192cb5cce33e8d23e0f1 Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.962679 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.962706 4862 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.962717 4862 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c005a75-9925-4968-bb0f-40c16c805d35-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.962812 4862 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.962824 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pskx\" (UniqueName: \"kubernetes.io/projected/5c005a75-9925-4968-bb0f-40c16c805d35-kube-api-access-2pskx\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.965722 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-config-data" (OuterVolumeSpecName: "config-data") pod "5c005a75-9925-4968-bb0f-40c16c805d35" (UID: "5c005a75-9925-4968-bb0f-40c16c805d35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:12 crc kubenswrapper[4862]: I1128 11:44:12.978546 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c005a75-9925-4968-bb0f-40c16c805d35" (UID: "5c005a75-9925-4968-bb0f-40c16c805d35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.051691 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-857784bb5b-4pc8q" event={"ID":"91afecb8-a07b-43e5-92e4-4e86cce7bcca","Type":"ContainerStarted","Data":"807c30986f85c37de51a7b74d0c8e012c3c5178a714e192cb5cce33e8d23e0f1"} Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.055786 4862 generic.go:334] "Generic (PLEG): container finished" podID="5c005a75-9925-4968-bb0f-40c16c805d35" containerID="07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d" exitCode=0 Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.055862 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerDied","Data":"07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d"} Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.055890 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c005a75-9925-4968-bb0f-40c16c805d35","Type":"ContainerDied","Data":"5dc6b2417038674a878c4d4565581172dd45cbff0efcbb6b441236e24c798b9d"} Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.055922 4862 scope.go:117] "RemoveContainer" containerID="97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.056169 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.068044 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.068081 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c005a75-9925-4968-bb0f-40c16c805d35-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.075385 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" event={"ID":"e4f23c53-e6f8-4aca-a1e4-969a58b70a38","Type":"ContainerStarted","Data":"b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6"} Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.075426 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" event={"ID":"e4f23c53-e6f8-4aca-a1e4-969a58b70a38","Type":"ContainerStarted","Data":"d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5"} Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.079249 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6df5c754df-6lvmr" event={"ID":"d5203b21-897c-4331-a878-1d5c949a3e75","Type":"ContainerStarted","Data":"31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7"} Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.079357 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6df5c754df-6lvmr" event={"ID":"d5203b21-897c-4331-a878-1d5c949a3e75","Type":"ContainerStarted","Data":"cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d"} Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.082942 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5jcj5" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.083977 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5jcj5" event={"ID":"b376a4e9-5456-4e8f-9c91-b222f38c30db","Type":"ContainerDied","Data":"33eb29d50e1b0895601a8eb8990f3270f01e8748228dbc7f014a85b73419dda1"} Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.083998 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33eb29d50e1b0895601a8eb8990f3270f01e8748228dbc7f014a85b73419dda1" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.084016 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.084088 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.084683 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.086726 4862 scope.go:117] "RemoveContainer" containerID="46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.114810 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" podStartSLOduration=2.613858915 podStartE2EDuration="4.114790765s" podCreationTimestamp="2025-11-28 11:44:09 +0000 UTC" firstStartedPulling="2025-11-28 11:44:10.37273496 +0000 UTC m=+1327.905248881" lastFinishedPulling="2025-11-28 11:44:11.87366681 +0000 UTC m=+1329.406180731" observedRunningTime="2025-11-28 11:44:13.092161642 +0000 UTC m=+1330.624675563" watchObservedRunningTime="2025-11-28 11:44:13.114790765 +0000 UTC m=+1330.647304686" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.120936 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6df5c754df-6lvmr" podStartSLOduration=2.392282813 podStartE2EDuration="4.120924652s" podCreationTimestamp="2025-11-28 11:44:09 +0000 UTC" firstStartedPulling="2025-11-28 11:44:10.1415315 +0000 UTC m=+1327.674045421" lastFinishedPulling="2025-11-28 11:44:11.870173339 +0000 UTC m=+1329.402687260" observedRunningTime="2025-11-28 11:44:13.112741986 +0000 UTC m=+1330.645255907" watchObservedRunningTime="2025-11-28 11:44:13.120924652 +0000 UTC m=+1330.653438573" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.145377 4862 scope.go:117] "RemoveContainer" containerID="07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.148171 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.166430 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.206721 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:13 crc kubenswrapper[4862]: E1128 11:44:13.207155 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="sg-core" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207167 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="sg-core" Nov 28 11:44:13 crc kubenswrapper[4862]: E1128 11:44:13.207198 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b376a4e9-5456-4e8f-9c91-b222f38c30db" containerName="cinder-db-sync" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207205 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b376a4e9-5456-4e8f-9c91-b222f38c30db" containerName="cinder-db-sync" Nov 28 11:44:13 crc kubenswrapper[4862]: E1128 11:44:13.207221 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="proxy-httpd" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207228 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="proxy-httpd" Nov 28 11:44:13 crc kubenswrapper[4862]: E1128 11:44:13.207241 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="ceilometer-notification-agent" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207247 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="ceilometer-notification-agent" Nov 28 11:44:13 crc kubenswrapper[4862]: E1128 11:44:13.207255 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="ceilometer-central-agent" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207261 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="ceilometer-central-agent" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207423 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b376a4e9-5456-4e8f-9c91-b222f38c30db" containerName="cinder-db-sync" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207437 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="proxy-httpd" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207448 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="sg-core" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207464 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="ceilometer-notification-agent" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.207481 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" containerName="ceilometer-central-agent" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.209082 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.226463 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.227575 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.227859 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.244596 4862 scope.go:117] "RemoveContainer" containerID="dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.265978 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.269043 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.274734 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.274763 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.274996 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ggp6j" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.277791 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.277922 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-scripts\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.278119 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.278233 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fwz9\" (UniqueName: \"kubernetes.io/projected/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-kube-api-access-8fwz9\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.278300 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-log-httpd\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.278368 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-run-httpd\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.278435 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-config-data\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.278689 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.294189 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.339778 4862 scope.go:117] "RemoveContainer" containerID="97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898" Nov 28 11:44:13 crc kubenswrapper[4862]: E1128 11:44:13.340309 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898\": container with ID starting with 97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898 not found: ID does not exist" containerID="97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.340346 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898"} err="failed to get container status \"97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898\": rpc error: code = NotFound desc = could not find container \"97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898\": container with ID starting with 97d56a1ae02a332cf9aacb985cfd804280194ad81e4de97e835543b639a9a898 not found: ID does not exist" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.340371 4862 scope.go:117] "RemoveContainer" containerID="46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce" Nov 28 11:44:13 crc kubenswrapper[4862]: E1128 11:44:13.340802 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce\": container with ID starting with 46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce not found: ID does not exist" containerID="46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.340822 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce"} err="failed to get container status \"46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce\": rpc error: code = NotFound desc = could not find container \"46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce\": container with ID starting with 46a0ca23cbf6c3f0b592d91e3333974eb126929c55573aefc5a4d1e29376a2ce not found: ID does not exist" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.340837 4862 scope.go:117] "RemoveContainer" containerID="07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d" Nov 28 11:44:13 crc kubenswrapper[4862]: E1128 11:44:13.341293 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d\": container with ID starting with 07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d not found: ID does not exist" containerID="07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.341316 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d"} err="failed to get container status \"07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d\": rpc error: code = NotFound desc = could not find container \"07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d\": container with ID starting with 07caf3fc8b6363fc2bc963ad02b092ec74636c3985589487ddde1922a1fb959d not found: ID does not exist" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.341331 4862 scope.go:117] "RemoveContainer" containerID="dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe" Nov 28 11:44:13 crc kubenswrapper[4862]: E1128 11:44:13.341534 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe\": container with ID starting with dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe not found: ID does not exist" containerID="dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.341557 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe"} err="failed to get container status \"dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe\": rpc error: code = NotFound desc = could not find container \"dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe\": container with ID starting with dd8af525393d27ad270bdad0940ebbcc9a4320b6f9f5a29f6d3b5f9bb7d79afe not found: ID does not exist" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.376997 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf96b7dc5-h7x89"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380556 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdxr2\" (UniqueName: \"kubernetes.io/projected/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-kube-api-access-gdxr2\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380632 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380659 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380681 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380710 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fwz9\" (UniqueName: \"kubernetes.io/projected/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-kube-api-access-8fwz9\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380737 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-log-httpd\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380756 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-run-httpd\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380772 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-config-data\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380800 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380828 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380865 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380894 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-scripts\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.380910 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.383217 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-log-httpd\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.383569 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-run-httpd\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.390495 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.390760 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.391805 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-scripts\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.395649 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-config-data\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.405789 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c55f6679-bknmm"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.407354 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.435837 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fwz9\" (UniqueName: \"kubernetes.io/projected/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-kube-api-access-8fwz9\") pod \"ceilometer-0\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.445372 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c55f6679-bknmm"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.482883 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.482942 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-config\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.484369 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.484508 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.484563 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-svc\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.484630 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-nb\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.484689 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmts9\" (UniqueName: \"kubernetes.io/projected/d61f3392-fcc7-42dd-8028-172d4f5e04b2-kube-api-access-vmts9\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.484800 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdxr2\" (UniqueName: \"kubernetes.io/projected/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-kube-api-access-gdxr2\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.484928 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.484965 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.485008 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-swift-storage-0\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.485056 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-sb\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.485235 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.499129 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.499517 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.499684 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.506059 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdxr2\" (UniqueName: \"kubernetes.io/projected/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-kube-api-access-gdxr2\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.507248 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.515539 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.517217 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.519256 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.536989 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.547443 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591282 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591342 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-swift-storage-0\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591365 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-sb\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591406 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-config\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591447 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/969e208b-74d1-4bd4-b4a6-db76d0b059fe-logs\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591470 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-svc\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591488 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zldws\" (UniqueName: \"kubernetes.io/projected/969e208b-74d1-4bd4-b4a6-db76d0b059fe-kube-api-access-zldws\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591508 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591523 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591545 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-nb\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591570 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmts9\" (UniqueName: \"kubernetes.io/projected/d61f3392-fcc7-42dd-8028-172d4f5e04b2-kube-api-access-vmts9\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591586 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-scripts\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.591610 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/969e208b-74d1-4bd4-b4a6-db76d0b059fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.592432 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-swift-storage-0\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.593024 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-nb\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.593987 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-config\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.594565 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-svc\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.594610 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.595049 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-sb\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.614142 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmts9\" (UniqueName: \"kubernetes.io/projected/d61f3392-fcc7-42dd-8028-172d4f5e04b2-kube-api-access-vmts9\") pod \"dnsmasq-dns-c55f6679-bknmm\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.702226 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/969e208b-74d1-4bd4-b4a6-db76d0b059fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.702326 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/969e208b-74d1-4bd4-b4a6-db76d0b059fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.702418 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.702546 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/969e208b-74d1-4bd4-b4a6-db76d0b059fe-logs\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.702598 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zldws\" (UniqueName: \"kubernetes.io/projected/969e208b-74d1-4bd4-b4a6-db76d0b059fe-kube-api-access-zldws\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.702623 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.702640 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.702709 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-scripts\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.703452 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/969e208b-74d1-4bd4-b4a6-db76d0b059fe-logs\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.706193 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.709597 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.709626 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.710803 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-scripts\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.749520 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zldws\" (UniqueName: \"kubernetes.io/projected/969e208b-74d1-4bd4-b4a6-db76d0b059fe-kube-api-access-zldws\") pod \"cinder-api-0\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " pod="openstack/cinder-api-0" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.863299 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:13 crc kubenswrapper[4862]: I1128 11:44:13.870650 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.002694 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.097748 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerStarted","Data":"1c97dfaf0d885762cecbf7531e0bdc9dc44a22faa74b1ef5ba0884a8cf0e8117"} Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.112787 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-857784bb5b-4pc8q" event={"ID":"91afecb8-a07b-43e5-92e4-4e86cce7bcca","Type":"ContainerStarted","Data":"773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92"} Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.112821 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-857784bb5b-4pc8q" event={"ID":"91afecb8-a07b-43e5-92e4-4e86cce7bcca","Type":"ContainerStarted","Data":"39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f"} Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.112983 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.136621 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.141506 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-857784bb5b-4pc8q" podStartSLOduration=3.141495244 podStartE2EDuration="3.141495244s" podCreationTimestamp="2025-11-28 11:44:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:14.139463016 +0000 UTC m=+1331.671976937" watchObservedRunningTime="2025-11-28 11:44:14.141495244 +0000 UTC m=+1331.674009165" Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.430742 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c55f6679-bknmm"] Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.516533 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:44:14 crc kubenswrapper[4862]: I1128 11:44:14.853512 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c005a75-9925-4968-bb0f-40c16c805d35" path="/var/lib/kubelet/pods/5c005a75-9925-4968-bb0f-40c16c805d35/volumes" Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.162839 4862 generic.go:334] "Generic (PLEG): container finished" podID="d61f3392-fcc7-42dd-8028-172d4f5e04b2" containerID="6a3cd0f4a9e484c02e4255709623e45cdfd370830a5817ea1e8c645d567544be" exitCode=0 Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.163040 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c55f6679-bknmm" event={"ID":"d61f3392-fcc7-42dd-8028-172d4f5e04b2","Type":"ContainerDied","Data":"6a3cd0f4a9e484c02e4255709623e45cdfd370830a5817ea1e8c645d567544be"} Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.163067 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c55f6679-bknmm" event={"ID":"d61f3392-fcc7-42dd-8028-172d4f5e04b2","Type":"ContainerStarted","Data":"bd5f19d9ff387922676603cab18fa3ece93be4cab7adce347b391e3bdcf2d8f1"} Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.171257 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"969e208b-74d1-4bd4-b4a6-db76d0b059fe","Type":"ContainerStarted","Data":"fc5d54e6d71a927b6e80f8967ce027e2238ce00f03e2d6a01083837226a80a8d"} Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.178809 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a","Type":"ContainerStarted","Data":"371201953cc39b149a78d8f5c894fcf8f3242bb4c9b0791db86c202b7186e458"} Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.184531 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerStarted","Data":"6c539e3b0541d5b88f8cfaddae50496a0032906b15d6865b7f041b3968483932"} Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.185826 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.186188 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" podUID="701596e6-38ae-4964-839d-44118647085e" containerName="dnsmasq-dns" containerID="cri-o://f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8" gracePeriod=10 Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.894676 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.957027 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-svc\") pod \"701596e6-38ae-4964-839d-44118647085e\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.957145 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-swift-storage-0\") pod \"701596e6-38ae-4964-839d-44118647085e\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.957169 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-sb\") pod \"701596e6-38ae-4964-839d-44118647085e\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.957201 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9dmc\" (UniqueName: \"kubernetes.io/projected/701596e6-38ae-4964-839d-44118647085e-kube-api-access-t9dmc\") pod \"701596e6-38ae-4964-839d-44118647085e\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.957311 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-nb\") pod \"701596e6-38ae-4964-839d-44118647085e\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.957337 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-config\") pod \"701596e6-38ae-4964-839d-44118647085e\" (UID: \"701596e6-38ae-4964-839d-44118647085e\") " Nov 28 11:44:15 crc kubenswrapper[4862]: I1128 11:44:15.970417 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/701596e6-38ae-4964-839d-44118647085e-kube-api-access-t9dmc" (OuterVolumeSpecName: "kube-api-access-t9dmc") pod "701596e6-38ae-4964-839d-44118647085e" (UID: "701596e6-38ae-4964-839d-44118647085e"). InnerVolumeSpecName "kube-api-access-t9dmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.023129 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "701596e6-38ae-4964-839d-44118647085e" (UID: "701596e6-38ae-4964-839d-44118647085e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.036648 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "701596e6-38ae-4964-839d-44118647085e" (UID: "701596e6-38ae-4964-839d-44118647085e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.051994 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "701596e6-38ae-4964-839d-44118647085e" (UID: "701596e6-38ae-4964-839d-44118647085e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.057997 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "701596e6-38ae-4964-839d-44118647085e" (UID: "701596e6-38ae-4964-839d-44118647085e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.059227 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.059247 4862 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.059256 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.059266 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9dmc\" (UniqueName: \"kubernetes.io/projected/701596e6-38ae-4964-839d-44118647085e-kube-api-access-t9dmc\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.059275 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.060324 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-config" (OuterVolumeSpecName: "config") pod "701596e6-38ae-4964-839d-44118647085e" (UID: "701596e6-38ae-4964-839d-44118647085e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.162431 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701596e6-38ae-4964-839d-44118647085e-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.201160 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.248898 4862 generic.go:334] "Generic (PLEG): container finished" podID="701596e6-38ae-4964-839d-44118647085e" containerID="f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8" exitCode=0 Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.248995 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" event={"ID":"701596e6-38ae-4964-839d-44118647085e","Type":"ContainerDied","Data":"f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8"} Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.249023 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" event={"ID":"701596e6-38ae-4964-839d-44118647085e","Type":"ContainerDied","Data":"8a696d0b6b75425a23786eed252c30d3ed5a29a92a1dfe64dd05657844f016bc"} Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.249040 4862 scope.go:117] "RemoveContainer" containerID="f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.249193 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf96b7dc5-h7x89" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.278503 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c55f6679-bknmm" event={"ID":"d61f3392-fcc7-42dd-8028-172d4f5e04b2","Type":"ContainerStarted","Data":"bde25c7a11d728f465b47e43d012c44bfe12d09c2b631dad13c8f797360f6927"} Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.281372 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.287969 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"969e208b-74d1-4bd4-b4a6-db76d0b059fe","Type":"ContainerStarted","Data":"9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc"} Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.298077 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf96b7dc5-h7x89"] Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.320147 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf96b7dc5-h7x89"] Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.320223 4862 scope.go:117] "RemoveContainer" containerID="d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.320436 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c55f6679-bknmm" podStartSLOduration=3.320416943 podStartE2EDuration="3.320416943s" podCreationTimestamp="2025-11-28 11:44:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:16.30956308 +0000 UTC m=+1333.842077001" watchObservedRunningTime="2025-11-28 11:44:16.320416943 +0000 UTC m=+1333.852930864" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.371133 4862 scope.go:117] "RemoveContainer" containerID="f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8" Nov 28 11:44:16 crc kubenswrapper[4862]: E1128 11:44:16.371677 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8\": container with ID starting with f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8 not found: ID does not exist" containerID="f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.371710 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8"} err="failed to get container status \"f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8\": rpc error: code = NotFound desc = could not find container \"f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8\": container with ID starting with f43bebfeb0d79ecd4b80bb846c128f4ad1dff22a616e370699c67b1b702f66a8 not found: ID does not exist" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.371733 4862 scope.go:117] "RemoveContainer" containerID="d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c" Nov 28 11:44:16 crc kubenswrapper[4862]: E1128 11:44:16.374916 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c\": container with ID starting with d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c not found: ID does not exist" containerID="d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.374952 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c"} err="failed to get container status \"d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c\": rpc error: code = NotFound desc = could not find container \"d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c\": container with ID starting with d4da4d3165df2de93a876cb2f3d6cf5fe9cee6c91230bf4d4621b7938fdbf31c not found: ID does not exist" Nov 28 11:44:16 crc kubenswrapper[4862]: I1128 11:44:16.850794 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="701596e6-38ae-4964-839d-44118647085e" path="/var/lib/kubelet/pods/701596e6-38ae-4964-839d-44118647085e/volumes" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.184061 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.298655 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerStarted","Data":"dd4d3eb043ce38a84be808f85db30333b93ef2b88d5dcd4c6bc3e1e69fe6661f"} Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.299041 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerStarted","Data":"53f1471876d0c8657cd1db9196437c738bda3153eb6235f76362ca2f52d69714"} Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.302115 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"969e208b-74d1-4bd4-b4a6-db76d0b059fe","Type":"ContainerStarted","Data":"e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1"} Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.302282 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerName="cinder-api-log" containerID="cri-o://9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc" gracePeriod=30 Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.302457 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerName="cinder-api" containerID="cri-o://e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1" gracePeriod=30 Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.302637 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.312705 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a","Type":"ContainerStarted","Data":"cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095"} Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.312875 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a","Type":"ContainerStarted","Data":"1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208"} Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.325752 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.325736355 podStartE2EDuration="4.325736355s" podCreationTimestamp="2025-11-28 11:44:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:17.320537925 +0000 UTC m=+1334.853051846" watchObservedRunningTime="2025-11-28 11:44:17.325736355 +0000 UTC m=+1334.858250276" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.354546 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.198765585 podStartE2EDuration="4.354524016s" podCreationTimestamp="2025-11-28 11:44:13 +0000 UTC" firstStartedPulling="2025-11-28 11:44:14.201243898 +0000 UTC m=+1331.733757819" lastFinishedPulling="2025-11-28 11:44:15.357002339 +0000 UTC m=+1332.889516250" observedRunningTime="2025-11-28 11:44:17.347974837 +0000 UTC m=+1334.880488758" watchObservedRunningTime="2025-11-28 11:44:17.354524016 +0000 UTC m=+1334.887037937" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.854260 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.922050 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data\") pod \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.922382 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/969e208b-74d1-4bd4-b4a6-db76d0b059fe-etc-machine-id\") pod \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.922485 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/969e208b-74d1-4bd4-b4a6-db76d0b059fe-logs\") pod \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.922569 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-scripts\") pod \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.922719 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data-custom\") pod \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.922802 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zldws\" (UniqueName: \"kubernetes.io/projected/969e208b-74d1-4bd4-b4a6-db76d0b059fe-kube-api-access-zldws\") pod \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.922882 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-combined-ca-bundle\") pod \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\" (UID: \"969e208b-74d1-4bd4-b4a6-db76d0b059fe\") " Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.922495 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/969e208b-74d1-4bd4-b4a6-db76d0b059fe-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "969e208b-74d1-4bd4-b4a6-db76d0b059fe" (UID: "969e208b-74d1-4bd4-b4a6-db76d0b059fe"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.923040 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/969e208b-74d1-4bd4-b4a6-db76d0b059fe-logs" (OuterVolumeSpecName: "logs") pod "969e208b-74d1-4bd4-b4a6-db76d0b059fe" (UID: "969e208b-74d1-4bd4-b4a6-db76d0b059fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.925522 4862 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/969e208b-74d1-4bd4-b4a6-db76d0b059fe-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.925559 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/969e208b-74d1-4bd4-b4a6-db76d0b059fe-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.929260 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-scripts" (OuterVolumeSpecName: "scripts") pod "969e208b-74d1-4bd4-b4a6-db76d0b059fe" (UID: "969e208b-74d1-4bd4-b4a6-db76d0b059fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.934204 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/969e208b-74d1-4bd4-b4a6-db76d0b059fe-kube-api-access-zldws" (OuterVolumeSpecName: "kube-api-access-zldws") pod "969e208b-74d1-4bd4-b4a6-db76d0b059fe" (UID: "969e208b-74d1-4bd4-b4a6-db76d0b059fe"). InnerVolumeSpecName "kube-api-access-zldws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.934222 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "969e208b-74d1-4bd4-b4a6-db76d0b059fe" (UID: "969e208b-74d1-4bd4-b4a6-db76d0b059fe"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:17 crc kubenswrapper[4862]: I1128 11:44:17.964000 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "969e208b-74d1-4bd4-b4a6-db76d0b059fe" (UID: "969e208b-74d1-4bd4-b4a6-db76d0b059fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.011245 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data" (OuterVolumeSpecName: "config-data") pod "969e208b-74d1-4bd4-b4a6-db76d0b059fe" (UID: "969e208b-74d1-4bd4-b4a6-db76d0b059fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.027384 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.027415 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.027427 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zldws\" (UniqueName: \"kubernetes.io/projected/969e208b-74d1-4bd4-b4a6-db76d0b059fe-kube-api-access-zldws\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.027436 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.027449 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/969e208b-74d1-4bd4-b4a6-db76d0b059fe-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.341819 4862 generic.go:334] "Generic (PLEG): container finished" podID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerID="e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1" exitCode=0 Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.341853 4862 generic.go:334] "Generic (PLEG): container finished" podID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerID="9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc" exitCode=143 Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.341870 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"969e208b-74d1-4bd4-b4a6-db76d0b059fe","Type":"ContainerDied","Data":"e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1"} Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.341897 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"969e208b-74d1-4bd4-b4a6-db76d0b059fe","Type":"ContainerDied","Data":"9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc"} Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.341908 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"969e208b-74d1-4bd4-b4a6-db76d0b059fe","Type":"ContainerDied","Data":"fc5d54e6d71a927b6e80f8967ce027e2238ce00f03e2d6a01083837226a80a8d"} Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.341922 4862 scope.go:117] "RemoveContainer" containerID="e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.341853 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.371710 4862 scope.go:117] "RemoveContainer" containerID="9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.497270 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.500247 4862 scope.go:117] "RemoveContainer" containerID="e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.504114 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:44:18 crc kubenswrapper[4862]: E1128 11:44:18.504820 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1\": container with ID starting with e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1 not found: ID does not exist" containerID="e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.504877 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1"} err="failed to get container status \"e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1\": rpc error: code = NotFound desc = could not find container \"e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1\": container with ID starting with e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1 not found: ID does not exist" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.504910 4862 scope.go:117] "RemoveContainer" containerID="9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc" Nov 28 11:44:18 crc kubenswrapper[4862]: E1128 11:44:18.506198 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc\": container with ID starting with 9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc not found: ID does not exist" containerID="9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.506241 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc"} err="failed to get container status \"9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc\": rpc error: code = NotFound desc = could not find container \"9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc\": container with ID starting with 9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc not found: ID does not exist" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.506262 4862 scope.go:117] "RemoveContainer" containerID="e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.507772 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1"} err="failed to get container status \"e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1\": rpc error: code = NotFound desc = could not find container \"e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1\": container with ID starting with e25fa7d0b17bce55e2d89dd5aa5ad4936e4038c08c54b2cc3a6b7f4c5b4457c1 not found: ID does not exist" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.507813 4862 scope.go:117] "RemoveContainer" containerID="9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.509367 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc"} err="failed to get container status \"9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc\": rpc error: code = NotFound desc = could not find container \"9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc\": container with ID starting with 9491ea34fe5ff6c1037d6d7eea77593c9c1c9774eab4328239f31c6c1ae12dcc not found: ID does not exist" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.520128 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:44:18 crc kubenswrapper[4862]: E1128 11:44:18.520927 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701596e6-38ae-4964-839d-44118647085e" containerName="dnsmasq-dns" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.521014 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="701596e6-38ae-4964-839d-44118647085e" containerName="dnsmasq-dns" Nov 28 11:44:18 crc kubenswrapper[4862]: E1128 11:44:18.521120 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerName="cinder-api" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.521200 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerName="cinder-api" Nov 28 11:44:18 crc kubenswrapper[4862]: E1128 11:44:18.521281 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701596e6-38ae-4964-839d-44118647085e" containerName="init" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.521349 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="701596e6-38ae-4964-839d-44118647085e" containerName="init" Nov 28 11:44:18 crc kubenswrapper[4862]: E1128 11:44:18.521428 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerName="cinder-api-log" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.521503 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerName="cinder-api-log" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.521807 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="701596e6-38ae-4964-839d-44118647085e" containerName="dnsmasq-dns" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.521891 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerName="cinder-api-log" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.521971 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" containerName="cinder-api" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.523003 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.529043 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.529229 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.531765 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.532410 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.595269 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.648184 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffc2928a-7858-4117-8786-723965b126f4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.648235 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.648283 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.648314 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27g6d\" (UniqueName: \"kubernetes.io/projected/ffc2928a-7858-4117-8786-723965b126f4-kube-api-access-27g6d\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.648371 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-scripts\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.648458 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.648522 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.648600 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data-custom\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.648676 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc2928a-7858-4117-8786-723965b126f4-logs\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.750232 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc2928a-7858-4117-8786-723965b126f4-logs\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.750316 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffc2928a-7858-4117-8786-723965b126f4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.750339 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.750377 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.750397 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27g6d\" (UniqueName: \"kubernetes.io/projected/ffc2928a-7858-4117-8786-723965b126f4-kube-api-access-27g6d\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.750439 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-scripts\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.750460 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.750501 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.750534 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data-custom\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.753217 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffc2928a-7858-4117-8786-723965b126f4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.753589 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc2928a-7858-4117-8786-723965b126f4-logs\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.757477 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data-custom\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.757837 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.758542 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.760541 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-scripts\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.760669 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.763396 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.773692 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27g6d\" (UniqueName: \"kubernetes.io/projected/ffc2928a-7858-4117-8786-723965b126f4-kube-api-access-27g6d\") pod \"cinder-api-0\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.849720 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="969e208b-74d1-4bd4-b4a6-db76d0b059fe" path="/var/lib/kubelet/pods/969e208b-74d1-4bd4-b4a6-db76d0b059fe/volumes" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.863868 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 11:44:18 crc kubenswrapper[4862]: I1128 11:44:18.902690 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:19 crc kubenswrapper[4862]: I1128 11:44:19.358832 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:44:19 crc kubenswrapper[4862]: I1128 11:44:19.359434 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerStarted","Data":"1923fadcaf39e74964ba821ad8477e6764522a15f42da72ff27733773c8c2ccb"} Nov 28 11:44:19 crc kubenswrapper[4862]: I1128 11:44:19.360151 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 11:44:19 crc kubenswrapper[4862]: I1128 11:44:19.406375 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.085497058 podStartE2EDuration="6.406351528s" podCreationTimestamp="2025-11-28 11:44:13 +0000 UTC" firstStartedPulling="2025-11-28 11:44:14.034072725 +0000 UTC m=+1331.566586646" lastFinishedPulling="2025-11-28 11:44:18.354927195 +0000 UTC m=+1335.887441116" observedRunningTime="2025-11-28 11:44:19.385245669 +0000 UTC m=+1336.917759600" watchObservedRunningTime="2025-11-28 11:44:19.406351528 +0000 UTC m=+1336.938865449" Nov 28 11:44:20 crc kubenswrapper[4862]: I1128 11:44:20.388491 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ffc2928a-7858-4117-8786-723965b126f4","Type":"ContainerStarted","Data":"72dd7e05b0e98629a3547aa5ced1d209fea4cbcd7cb7dc51836836e79ae4b826"} Nov 28 11:44:20 crc kubenswrapper[4862]: I1128 11:44:20.389086 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ffc2928a-7858-4117-8786-723965b126f4","Type":"ContainerStarted","Data":"a1504d4191ee33f529cd3c01feba8563b3fed4553cd518928314cae338b4ede1"} Nov 28 11:44:21 crc kubenswrapper[4862]: I1128 11:44:21.402601 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ffc2928a-7858-4117-8786-723965b126f4","Type":"ContainerStarted","Data":"2732ee7a075e6578c8913e90a93812b8f69feb263ac51f3983a02e4f5cb85ed0"} Nov 28 11:44:21 crc kubenswrapper[4862]: I1128 11:44:21.402944 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 28 11:44:21 crc kubenswrapper[4862]: I1128 11:44:21.435935 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.435910889 podStartE2EDuration="3.435910889s" podCreationTimestamp="2025-11-28 11:44:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:21.434322473 +0000 UTC m=+1338.966836394" watchObservedRunningTime="2025-11-28 11:44:21.435910889 +0000 UTC m=+1338.968424800" Nov 28 11:44:22 crc kubenswrapper[4862]: I1128 11:44:22.341619 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:44:23 crc kubenswrapper[4862]: I1128 11:44:23.822965 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:23 crc kubenswrapper[4862]: I1128 11:44:23.834346 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:44:23 crc kubenswrapper[4862]: I1128 11:44:23.865297 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:44:23 crc kubenswrapper[4862]: I1128 11:44:23.910320 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 28 11:44:23 crc kubenswrapper[4862]: I1128 11:44:23.935963 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-599f8c654b-clr4b"] Nov 28 11:44:23 crc kubenswrapper[4862]: I1128 11:44:23.936184 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-599f8c654b-clr4b" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api-log" containerID="cri-o://6b20364c8c922514b37f40471f7578b3c0622702bd785eb84a5903eb3f50eb68" gracePeriod=30 Nov 28 11:44:23 crc kubenswrapper[4862]: I1128 11:44:23.936314 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-599f8c654b-clr4b" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api" containerID="cri-o://adb727d122542042ce3d39254042b51e58e9fb18c8bdcf839eb1e8cf626fd5cd" gracePeriod=30 Nov 28 11:44:23 crc kubenswrapper[4862]: I1128 11:44:23.986263 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.010598 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f8dc44d89-t4bpq"] Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.010848 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" podUID="514236cd-f07b-41fd-8f58-f45c800642e6" containerName="dnsmasq-dns" containerID="cri-o://9b6db23f3e4130aa66413b92d4a49a64186b4cf841ad5c92cd550d02fbb0bb42" gracePeriod=10 Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.457900 4862 generic.go:334] "Generic (PLEG): container finished" podID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerID="6b20364c8c922514b37f40471f7578b3c0622702bd785eb84a5903eb3f50eb68" exitCode=143 Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.458396 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-599f8c654b-clr4b" event={"ID":"e0f90517-7935-4112-bcc9-07ac2e3cd2f8","Type":"ContainerDied","Data":"6b20364c8c922514b37f40471f7578b3c0622702bd785eb84a5903eb3f50eb68"} Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.460767 4862 generic.go:334] "Generic (PLEG): container finished" podID="514236cd-f07b-41fd-8f58-f45c800642e6" containerID="9b6db23f3e4130aa66413b92d4a49a64186b4cf841ad5c92cd550d02fbb0bb42" exitCode=0 Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.461996 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" event={"ID":"514236cd-f07b-41fd-8f58-f45c800642e6","Type":"ContainerDied","Data":"9b6db23f3e4130aa66413b92d4a49a64186b4cf841ad5c92cd550d02fbb0bb42"} Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.462899 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerName="probe" containerID="cri-o://cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095" gracePeriod=30 Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.462968 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerName="cinder-scheduler" containerID="cri-o://1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208" gracePeriod=30 Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.537320 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.664537 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-svc\") pod \"514236cd-f07b-41fd-8f58-f45c800642e6\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.664682 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-config\") pod \"514236cd-f07b-41fd-8f58-f45c800642e6\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.664705 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-nb\") pod \"514236cd-f07b-41fd-8f58-f45c800642e6\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.664774 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-sb\") pod \"514236cd-f07b-41fd-8f58-f45c800642e6\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.664836 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-swift-storage-0\") pod \"514236cd-f07b-41fd-8f58-f45c800642e6\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.664869 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c5kv\" (UniqueName: \"kubernetes.io/projected/514236cd-f07b-41fd-8f58-f45c800642e6-kube-api-access-2c5kv\") pod \"514236cd-f07b-41fd-8f58-f45c800642e6\" (UID: \"514236cd-f07b-41fd-8f58-f45c800642e6\") " Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.683297 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/514236cd-f07b-41fd-8f58-f45c800642e6-kube-api-access-2c5kv" (OuterVolumeSpecName: "kube-api-access-2c5kv") pod "514236cd-f07b-41fd-8f58-f45c800642e6" (UID: "514236cd-f07b-41fd-8f58-f45c800642e6"). InnerVolumeSpecName "kube-api-access-2c5kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.720404 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-config" (OuterVolumeSpecName: "config") pod "514236cd-f07b-41fd-8f58-f45c800642e6" (UID: "514236cd-f07b-41fd-8f58-f45c800642e6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.720721 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "514236cd-f07b-41fd-8f58-f45c800642e6" (UID: "514236cd-f07b-41fd-8f58-f45c800642e6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.731168 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "514236cd-f07b-41fd-8f58-f45c800642e6" (UID: "514236cd-f07b-41fd-8f58-f45c800642e6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.734479 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "514236cd-f07b-41fd-8f58-f45c800642e6" (UID: "514236cd-f07b-41fd-8f58-f45c800642e6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.742787 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "514236cd-f07b-41fd-8f58-f45c800642e6" (UID: "514236cd-f07b-41fd-8f58-f45c800642e6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.766501 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.766530 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.766543 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.766551 4862 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.766562 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c5kv\" (UniqueName: \"kubernetes.io/projected/514236cd-f07b-41fd-8f58-f45c800642e6-kube-api-access-2c5kv\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:24 crc kubenswrapper[4862]: I1128 11:44:24.766570 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/514236cd-f07b-41fd-8f58-f45c800642e6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.158309 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.270556 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6574c94c76-hq6jb"] Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.279542 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6574c94c76-hq6jb" podUID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerName="neutron-api" containerID="cri-o://807efefdc1c9165b39db99614c63e8fc78929a6ac95c8615b88dc151742fe58e" gracePeriod=30 Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.280150 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6574c94c76-hq6jb" podUID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerName="neutron-httpd" containerID="cri-o://c12815bdfd56949b92fac3d59058f4d61ac424748d6ad9871cb19fcac768ffac" gracePeriod=30 Nov 28 11:44:25 crc kubenswrapper[4862]: E1128 11:44:25.468368 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfe13255_9ab0_4dcd_baac_83d4ceb6a59a.slice/crio-conmon-cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095.scope\": RecentStats: unable to find data in memory cache]" Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.473240 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" event={"ID":"514236cd-f07b-41fd-8f58-f45c800642e6","Type":"ContainerDied","Data":"9446f531d7af9edc09a967e9e43ec88821fcf9d8f8716a6984f8a3f5ea1a06ce"} Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.473296 4862 scope.go:117] "RemoveContainer" containerID="9b6db23f3e4130aa66413b92d4a49a64186b4cf841ad5c92cd550d02fbb0bb42" Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.473347 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8dc44d89-t4bpq" Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.583437 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f8dc44d89-t4bpq"] Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.591208 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f8dc44d89-t4bpq"] Nov 28 11:44:25 crc kubenswrapper[4862]: I1128 11:44:25.597173 4862 scope.go:117] "RemoveContainer" containerID="f8f78fa6bc3eead90a890f3dd5d1f3fb7b5524a7998ad40d2954ed2bd92a945c" Nov 28 11:44:26 crc kubenswrapper[4862]: I1128 11:44:26.488260 4862 generic.go:334] "Generic (PLEG): container finished" podID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerID="c12815bdfd56949b92fac3d59058f4d61ac424748d6ad9871cb19fcac768ffac" exitCode=0 Nov 28 11:44:26 crc kubenswrapper[4862]: I1128 11:44:26.488357 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6574c94c76-hq6jb" event={"ID":"225d091e-4cad-444e-aa3c-dcdf65dff654","Type":"ContainerDied","Data":"c12815bdfd56949b92fac3d59058f4d61ac424748d6ad9871cb19fcac768ffac"} Nov 28 11:44:26 crc kubenswrapper[4862]: I1128 11:44:26.493371 4862 generic.go:334] "Generic (PLEG): container finished" podID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerID="cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095" exitCode=0 Nov 28 11:44:26 crc kubenswrapper[4862]: I1128 11:44:26.493421 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a","Type":"ContainerDied","Data":"cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095"} Nov 28 11:44:26 crc kubenswrapper[4862]: I1128 11:44:26.850290 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="514236cd-f07b-41fd-8f58-f45c800642e6" path="/var/lib/kubelet/pods/514236cd-f07b-41fd-8f58-f45c800642e6/volumes" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.108227 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-599f8c654b-clr4b" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:47856->10.217.0.156:9311: read: connection reset by peer" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.108286 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-599f8c654b-clr4b" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:47862->10.217.0.156:9311: read: connection reset by peer" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.502343 4862 generic.go:334] "Generic (PLEG): container finished" podID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerID="adb727d122542042ce3d39254042b51e58e9fb18c8bdcf839eb1e8cf626fd5cd" exitCode=0 Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.502628 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-599f8c654b-clr4b" event={"ID":"e0f90517-7935-4112-bcc9-07ac2e3cd2f8","Type":"ContainerDied","Data":"adb727d122542042ce3d39254042b51e58e9fb18c8bdcf839eb1e8cf626fd5cd"} Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.566677 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.567801 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.648186 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.725336 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data-custom\") pod \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.725416 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7pnf\" (UniqueName: \"kubernetes.io/projected/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-kube-api-access-v7pnf\") pod \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.725566 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-combined-ca-bundle\") pod \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.725608 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data\") pod \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.725681 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-logs\") pod \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\" (UID: \"e0f90517-7935-4112-bcc9-07ac2e3cd2f8\") " Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.726570 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-logs" (OuterVolumeSpecName: "logs") pod "e0f90517-7935-4112-bcc9-07ac2e3cd2f8" (UID: "e0f90517-7935-4112-bcc9-07ac2e3cd2f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.749300 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e0f90517-7935-4112-bcc9-07ac2e3cd2f8" (UID: "e0f90517-7935-4112-bcc9-07ac2e3cd2f8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.749353 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-kube-api-access-v7pnf" (OuterVolumeSpecName: "kube-api-access-v7pnf") pod "e0f90517-7935-4112-bcc9-07ac2e3cd2f8" (UID: "e0f90517-7935-4112-bcc9-07ac2e3cd2f8"). InnerVolumeSpecName "kube-api-access-v7pnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.783211 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data" (OuterVolumeSpecName: "config-data") pod "e0f90517-7935-4112-bcc9-07ac2e3cd2f8" (UID: "e0f90517-7935-4112-bcc9-07ac2e3cd2f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.799962 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0f90517-7935-4112-bcc9-07ac2e3cd2f8" (UID: "e0f90517-7935-4112-bcc9-07ac2e3cd2f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.828386 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.828414 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.828423 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.828433 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7pnf\" (UniqueName: \"kubernetes.io/projected/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-kube-api-access-v7pnf\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:27 crc kubenswrapper[4862]: I1128 11:44:27.828441 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f90517-7935-4112-bcc9-07ac2e3cd2f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.172659 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.239043 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-scripts\") pod \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.239113 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data-custom\") pod \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.239226 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data\") pod \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.239425 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-combined-ca-bundle\") pod \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.239702 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-etc-machine-id\") pod \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.239789 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" (UID: "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.239852 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdxr2\" (UniqueName: \"kubernetes.io/projected/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-kube-api-access-gdxr2\") pod \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\" (UID: \"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a\") " Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.240422 4862 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.251354 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-scripts" (OuterVolumeSpecName: "scripts") pod "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" (UID: "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.251543 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-kube-api-access-gdxr2" (OuterVolumeSpecName: "kube-api-access-gdxr2") pod "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" (UID: "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a"). InnerVolumeSpecName "kube-api-access-gdxr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.261958 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" (UID: "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.291668 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" (UID: "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.327023 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data" (OuterVolumeSpecName: "config-data") pod "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" (UID: "bfe13255-9ab0-4dcd-baac-83d4ceb6a59a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.342536 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.342571 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.342585 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.342597 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.342610 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdxr2\" (UniqueName: \"kubernetes.io/projected/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a-kube-api-access-gdxr2\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.577398 4862 generic.go:334] "Generic (PLEG): container finished" podID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerID="1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208" exitCode=0 Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.577454 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a","Type":"ContainerDied","Data":"1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208"} Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.577496 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe13255-9ab0-4dcd-baac-83d4ceb6a59a","Type":"ContainerDied","Data":"371201953cc39b149a78d8f5c894fcf8f3242bb4c9b0791db86c202b7186e458"} Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.577513 4862 scope.go:117] "RemoveContainer" containerID="cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.577432 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.588514 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-599f8c654b-clr4b" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.589147 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-599f8c654b-clr4b" event={"ID":"e0f90517-7935-4112-bcc9-07ac2e3cd2f8","Type":"ContainerDied","Data":"33e7682ec7f5c765a4e00f3fa3d4a98d1c6bd42fc1113a9e2c1bd7b76b348277"} Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.607880 4862 scope.go:117] "RemoveContainer" containerID="1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.621394 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-599f8c654b-clr4b"] Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.628513 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-599f8c654b-clr4b"] Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.638193 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.645342 4862 scope.go:117] "RemoveContainer" containerID="cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095" Nov 28 11:44:28 crc kubenswrapper[4862]: E1128 11:44:28.647471 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095\": container with ID starting with cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095 not found: ID does not exist" containerID="cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.647528 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095"} err="failed to get container status \"cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095\": rpc error: code = NotFound desc = could not find container \"cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095\": container with ID starting with cd733bc5bd1d4036a16f6f2dc392b3564aec21e7d39e4ea7b5a271bc9c0c9095 not found: ID does not exist" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.647559 4862 scope.go:117] "RemoveContainer" containerID="1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208" Nov 28 11:44:28 crc kubenswrapper[4862]: E1128 11:44:28.647938 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208\": container with ID starting with 1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208 not found: ID does not exist" containerID="1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.647972 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208"} err="failed to get container status \"1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208\": rpc error: code = NotFound desc = could not find container \"1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208\": container with ID starting with 1451476d64a581abda2b112e96c5378978c8ad4d00a5c5e18aa06cca63027208 not found: ID does not exist" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.647996 4862 scope.go:117] "RemoveContainer" containerID="adb727d122542042ce3d39254042b51e58e9fb18c8bdcf839eb1e8cf626fd5cd" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.665414 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.673242 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:44:28 crc kubenswrapper[4862]: E1128 11:44:28.673744 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerName="cinder-scheduler" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.673759 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerName="cinder-scheduler" Nov 28 11:44:28 crc kubenswrapper[4862]: E1128 11:44:28.673783 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="514236cd-f07b-41fd-8f58-f45c800642e6" containerName="dnsmasq-dns" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.673791 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="514236cd-f07b-41fd-8f58-f45c800642e6" containerName="dnsmasq-dns" Nov 28 11:44:28 crc kubenswrapper[4862]: E1128 11:44:28.673815 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="514236cd-f07b-41fd-8f58-f45c800642e6" containerName="init" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.673832 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="514236cd-f07b-41fd-8f58-f45c800642e6" containerName="init" Nov 28 11:44:28 crc kubenswrapper[4862]: E1128 11:44:28.673848 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerName="probe" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.673856 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerName="probe" Nov 28 11:44:28 crc kubenswrapper[4862]: E1128 11:44:28.673872 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.673882 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api" Nov 28 11:44:28 crc kubenswrapper[4862]: E1128 11:44:28.673900 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api-log" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.673908 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api-log" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.674151 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api-log" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.674176 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="514236cd-f07b-41fd-8f58-f45c800642e6" containerName="dnsmasq-dns" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.674192 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" containerName="barbican-api" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.674211 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerName="cinder-scheduler" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.674222 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" containerName="probe" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.675461 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.688570 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.693139 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.695964 4862 scope.go:117] "RemoveContainer" containerID="6b20364c8c922514b37f40471f7578b3c0622702bd785eb84a5903eb3f50eb68" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.750596 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.750834 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/989edff9-2d95-4839-9926-f5ee7dc65d19-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.750901 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.751578 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pccc2\" (UniqueName: \"kubernetes.io/projected/989edff9-2d95-4839-9926-f5ee7dc65d19-kube-api-access-pccc2\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.751639 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-scripts\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.751747 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.853910 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.854337 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.854435 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/989edff9-2d95-4839-9926-f5ee7dc65d19-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.854499 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/989edff9-2d95-4839-9926-f5ee7dc65d19-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.854509 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.854575 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pccc2\" (UniqueName: \"kubernetes.io/projected/989edff9-2d95-4839-9926-f5ee7dc65d19-kube-api-access-pccc2\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.854645 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-scripts\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.861938 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-scripts\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.862243 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.866859 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.868613 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.872286 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe13255-9ab0-4dcd-baac-83d4ceb6a59a" path="/var/lib/kubelet/pods/bfe13255-9ab0-4dcd-baac-83d4ceb6a59a/volumes" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.873350 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f90517-7935-4112-bcc9-07ac2e3cd2f8" path="/var/lib/kubelet/pods/e0f90517-7935-4112-bcc9-07ac2e3cd2f8/volumes" Nov 28 11:44:28 crc kubenswrapper[4862]: I1128 11:44:28.876160 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pccc2\" (UniqueName: \"kubernetes.io/projected/989edff9-2d95-4839-9926-f5ee7dc65d19-kube-api-access-pccc2\") pod \"cinder-scheduler-0\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " pod="openstack/cinder-scheduler-0" Nov 28 11:44:29 crc kubenswrapper[4862]: I1128 11:44:29.003497 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 11:44:29 crc kubenswrapper[4862]: I1128 11:44:29.267689 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:44:29 crc kubenswrapper[4862]: I1128 11:44:29.599861 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"989edff9-2d95-4839-9926-f5ee7dc65d19","Type":"ContainerStarted","Data":"8e6ad7cefc6df96185f6ce389da91df490b069298a8999fae01055de3567d518"} Nov 28 11:44:30 crc kubenswrapper[4862]: I1128 11:44:30.619165 4862 generic.go:334] "Generic (PLEG): container finished" podID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerID="807efefdc1c9165b39db99614c63e8fc78929a6ac95c8615b88dc151742fe58e" exitCode=0 Nov 28 11:44:30 crc kubenswrapper[4862]: I1128 11:44:30.619210 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6574c94c76-hq6jb" event={"ID":"225d091e-4cad-444e-aa3c-dcdf65dff654","Type":"ContainerDied","Data":"807efefdc1c9165b39db99614c63e8fc78929a6ac95c8615b88dc151742fe58e"} Nov 28 11:44:30 crc kubenswrapper[4862]: I1128 11:44:30.621612 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"989edff9-2d95-4839-9926-f5ee7dc65d19","Type":"ContainerStarted","Data":"5d77e5ca098b36fcc98d53cfa5e572df42e3b1667b2235d6c42ff31fe0c66c1f"} Nov 28 11:44:30 crc kubenswrapper[4862]: I1128 11:44:30.621635 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"989edff9-2d95-4839-9926-f5ee7dc65d19","Type":"ContainerStarted","Data":"8031be805d28d00e73f2d74cac7b5638f7feeaa4fd5d9e786af7a60f92caf4cc"} Nov 28 11:44:30 crc kubenswrapper[4862]: I1128 11:44:30.754239 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 28 11:44:30 crc kubenswrapper[4862]: I1128 11:44:30.778810 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.778788368 podStartE2EDuration="2.778788368s" podCreationTimestamp="2025-11-28 11:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:30.650469736 +0000 UTC m=+1348.182983667" watchObservedRunningTime="2025-11-28 11:44:30.778788368 +0000 UTC m=+1348.311302289" Nov 28 11:44:30 crc kubenswrapper[4862]: I1128 11:44:30.987027 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.109008 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-config\") pod \"225d091e-4cad-444e-aa3c-dcdf65dff654\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.109187 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsfsf\" (UniqueName: \"kubernetes.io/projected/225d091e-4cad-444e-aa3c-dcdf65dff654-kube-api-access-lsfsf\") pod \"225d091e-4cad-444e-aa3c-dcdf65dff654\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.109299 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-httpd-config\") pod \"225d091e-4cad-444e-aa3c-dcdf65dff654\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.109381 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-ovndb-tls-certs\") pod \"225d091e-4cad-444e-aa3c-dcdf65dff654\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.109451 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-combined-ca-bundle\") pod \"225d091e-4cad-444e-aa3c-dcdf65dff654\" (UID: \"225d091e-4cad-444e-aa3c-dcdf65dff654\") " Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.115989 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/225d091e-4cad-444e-aa3c-dcdf65dff654-kube-api-access-lsfsf" (OuterVolumeSpecName: "kube-api-access-lsfsf") pod "225d091e-4cad-444e-aa3c-dcdf65dff654" (UID: "225d091e-4cad-444e-aa3c-dcdf65dff654"). InnerVolumeSpecName "kube-api-access-lsfsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.116213 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "225d091e-4cad-444e-aa3c-dcdf65dff654" (UID: "225d091e-4cad-444e-aa3c-dcdf65dff654"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.198337 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "225d091e-4cad-444e-aa3c-dcdf65dff654" (UID: "225d091e-4cad-444e-aa3c-dcdf65dff654"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.211138 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsfsf\" (UniqueName: \"kubernetes.io/projected/225d091e-4cad-444e-aa3c-dcdf65dff654-kube-api-access-lsfsf\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.211171 4862 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.211180 4862 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.212159 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "225d091e-4cad-444e-aa3c-dcdf65dff654" (UID: "225d091e-4cad-444e-aa3c-dcdf65dff654"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.215002 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-config" (OuterVolumeSpecName: "config") pod "225d091e-4cad-444e-aa3c-dcdf65dff654" (UID: "225d091e-4cad-444e-aa3c-dcdf65dff654"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.313343 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.313376 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/225d091e-4cad-444e-aa3c-dcdf65dff654-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.640038 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6574c94c76-hq6jb" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.640052 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6574c94c76-hq6jb" event={"ID":"225d091e-4cad-444e-aa3c-dcdf65dff654","Type":"ContainerDied","Data":"210790da95d404a97863f733039f054b289df93ea24cf3b4f3c90ae8a4b23403"} Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.640773 4862 scope.go:117] "RemoveContainer" containerID="c12815bdfd56949b92fac3d59058f4d61ac424748d6ad9871cb19fcac768ffac" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.681953 4862 scope.go:117] "RemoveContainer" containerID="807efefdc1c9165b39db99614c63e8fc78929a6ac95c8615b88dc151742fe58e" Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.704994 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6574c94c76-hq6jb"] Nov 28 11:44:31 crc kubenswrapper[4862]: I1128 11:44:31.715069 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6574c94c76-hq6jb"] Nov 28 11:44:32 crc kubenswrapper[4862]: I1128 11:44:32.618487 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:44:32 crc kubenswrapper[4862]: I1128 11:44:32.857184 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="225d091e-4cad-444e-aa3c-dcdf65dff654" path="/var/lib/kubelet/pods/225d091e-4cad-444e-aa3c-dcdf65dff654/volumes" Nov 28 11:44:34 crc kubenswrapper[4862]: I1128 11:44:34.005402 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 28 11:44:34 crc kubenswrapper[4862]: I1128 11:44:34.244289 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.857698 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 11:44:35 crc kubenswrapper[4862]: E1128 11:44:35.858446 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerName="neutron-api" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.858460 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerName="neutron-api" Nov 28 11:44:35 crc kubenswrapper[4862]: E1128 11:44:35.858470 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerName="neutron-httpd" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.858476 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerName="neutron-httpd" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.858677 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerName="neutron-httpd" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.858690 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="225d091e-4cad-444e-aa3c-dcdf65dff654" containerName="neutron-api" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.859324 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.861814 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.862573 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.872232 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-jlqwf" Nov 28 11:44:35 crc kubenswrapper[4862]: I1128 11:44:35.875341 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.006485 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbnlt\" (UniqueName: \"kubernetes.io/projected/96e2f99b-d661-43af-8dc0-f762a14f9b3f-kube-api-access-cbnlt\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.006533 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.006577 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.006600 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config-secret\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.108305 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.108570 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config-secret\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.108688 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbnlt\" (UniqueName: \"kubernetes.io/projected/96e2f99b-d661-43af-8dc0-f762a14f9b3f-kube-api-access-cbnlt\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.108708 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.109465 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.115658 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config-secret\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.115845 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.130549 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbnlt\" (UniqueName: \"kubernetes.io/projected/96e2f99b-d661-43af-8dc0-f762a14f9b3f-kube-api-access-cbnlt\") pod \"openstackclient\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.176366 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.179329 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.185339 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.252210 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.253472 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.260845 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.312900 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-combined-ca-bundle\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.312956 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.313011 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxh5z\" (UniqueName: \"kubernetes.io/projected/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-kube-api-access-pxh5z\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.313029 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config-secret\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: E1128 11:44:36.329469 4862 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 28 11:44:36 crc kubenswrapper[4862]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_96e2f99b-d661-43af-8dc0-f762a14f9b3f_0(39986e199e9d853178e6d1fb6d5f2d2bb9d092fdb73d9764ff5962faab3b6e24): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"39986e199e9d853178e6d1fb6d5f2d2bb9d092fdb73d9764ff5962faab3b6e24" Netns:"/var/run/netns/ea47328d-23c8-4126-9e01-512a92acb2f3" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=39986e199e9d853178e6d1fb6d5f2d2bb9d092fdb73d9764ff5962faab3b6e24;K8S_POD_UID=96e2f99b-d661-43af-8dc0-f762a14f9b3f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/96e2f99b-d661-43af-8dc0-f762a14f9b3f]: expected pod UID "96e2f99b-d661-43af-8dc0-f762a14f9b3f" but got "23891b0e-3fc7-4d2f-9abf-96b57b09ca01" from Kube API Nov 28 11:44:36 crc kubenswrapper[4862]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 28 11:44:36 crc kubenswrapper[4862]: > Nov 28 11:44:36 crc kubenswrapper[4862]: E1128 11:44:36.329539 4862 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 28 11:44:36 crc kubenswrapper[4862]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_96e2f99b-d661-43af-8dc0-f762a14f9b3f_0(39986e199e9d853178e6d1fb6d5f2d2bb9d092fdb73d9764ff5962faab3b6e24): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"39986e199e9d853178e6d1fb6d5f2d2bb9d092fdb73d9764ff5962faab3b6e24" Netns:"/var/run/netns/ea47328d-23c8-4126-9e01-512a92acb2f3" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=39986e199e9d853178e6d1fb6d5f2d2bb9d092fdb73d9764ff5962faab3b6e24;K8S_POD_UID=96e2f99b-d661-43af-8dc0-f762a14f9b3f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/96e2f99b-d661-43af-8dc0-f762a14f9b3f]: expected pod UID "96e2f99b-d661-43af-8dc0-f762a14f9b3f" but got "23891b0e-3fc7-4d2f-9abf-96b57b09ca01" from Kube API Nov 28 11:44:36 crc kubenswrapper[4862]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 28 11:44:36 crc kubenswrapper[4862]: > pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.414610 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.414732 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxh5z\" (UniqueName: \"kubernetes.io/projected/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-kube-api-access-pxh5z\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.414768 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config-secret\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.414870 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-combined-ca-bundle\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.415450 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.419158 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config-secret\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.419637 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-combined-ca-bundle\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.434542 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxh5z\" (UniqueName: \"kubernetes.io/projected/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-kube-api-access-pxh5z\") pod \"openstackclient\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.599759 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.720008 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.729726 4862 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="96e2f99b-d661-43af-8dc0-f762a14f9b3f" podUID="23891b0e-3fc7-4d2f-9abf-96b57b09ca01" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.742519 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.822693 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbnlt\" (UniqueName: \"kubernetes.io/projected/96e2f99b-d661-43af-8dc0-f762a14f9b3f-kube-api-access-cbnlt\") pod \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.822732 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config\") pod \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.822827 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config-secret\") pod \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.822990 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-combined-ca-bundle\") pod \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\" (UID: \"96e2f99b-d661-43af-8dc0-f762a14f9b3f\") " Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.824834 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "96e2f99b-d661-43af-8dc0-f762a14f9b3f" (UID: "96e2f99b-d661-43af-8dc0-f762a14f9b3f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.831571 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "96e2f99b-d661-43af-8dc0-f762a14f9b3f" (UID: "96e2f99b-d661-43af-8dc0-f762a14f9b3f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.831771 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96e2f99b-d661-43af-8dc0-f762a14f9b3f" (UID: "96e2f99b-d661-43af-8dc0-f762a14f9b3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.834780 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e2f99b-d661-43af-8dc0-f762a14f9b3f-kube-api-access-cbnlt" (OuterVolumeSpecName: "kube-api-access-cbnlt") pod "96e2f99b-d661-43af-8dc0-f762a14f9b3f" (UID: "96e2f99b-d661-43af-8dc0-f762a14f9b3f"). InnerVolumeSpecName "kube-api-access-cbnlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.853056 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96e2f99b-d661-43af-8dc0-f762a14f9b3f" path="/var/lib/kubelet/pods/96e2f99b-d661-43af-8dc0-f762a14f9b3f/volumes" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.925274 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.925300 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbnlt\" (UniqueName: \"kubernetes.io/projected/96e2f99b-d661-43af-8dc0-f762a14f9b3f-kube-api-access-cbnlt\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.925311 4862 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:36 crc kubenswrapper[4862]: I1128 11:44:36.925321 4862 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96e2f99b-d661-43af-8dc0-f762a14f9b3f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:37 crc kubenswrapper[4862]: I1128 11:44:37.055647 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 11:44:37 crc kubenswrapper[4862]: W1128 11:44:37.059314 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23891b0e_3fc7_4d2f_9abf_96b57b09ca01.slice/crio-4fb8355290c76f1b9e840956935cb4162e34d4c10e8ededf078868115fa3d98d WatchSource:0}: Error finding container 4fb8355290c76f1b9e840956935cb4162e34d4c10e8ededf078868115fa3d98d: Status 404 returned error can't find the container with id 4fb8355290c76f1b9e840956935cb4162e34d4c10e8ededf078868115fa3d98d Nov 28 11:44:37 crc kubenswrapper[4862]: I1128 11:44:37.730567 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"23891b0e-3fc7-4d2f-9abf-96b57b09ca01","Type":"ContainerStarted","Data":"4fb8355290c76f1b9e840956935cb4162e34d4c10e8ededf078868115fa3d98d"} Nov 28 11:44:37 crc kubenswrapper[4862]: I1128 11:44:37.730586 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:44:37 crc kubenswrapper[4862]: I1128 11:44:37.737633 4862 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="96e2f99b-d661-43af-8dc0-f762a14f9b3f" podUID="23891b0e-3fc7-4d2f-9abf-96b57b09ca01" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.508195 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-689b7bfc65-h98mj"] Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.511281 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.517291 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.519961 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.520715 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-689b7bfc65-h98mj"] Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.521824 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.577935 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-public-tls-certs\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.577984 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-run-httpd\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.578009 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-internal-tls-certs\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.578039 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svqnq\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-kube-api-access-svqnq\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.578060 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-log-httpd\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.578114 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-etc-swift\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.578151 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-config-data\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.578175 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-combined-ca-bundle\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.679823 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-public-tls-certs\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.680194 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-run-httpd\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.680225 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-internal-tls-certs\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.680254 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svqnq\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-kube-api-access-svqnq\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.680273 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-log-httpd\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.680320 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-etc-swift\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.680354 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-config-data\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.680378 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-combined-ca-bundle\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.680636 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-run-httpd\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.680787 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-log-httpd\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.688363 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-internal-tls-certs\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.691105 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-public-tls-certs\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.708031 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-combined-ca-bundle\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.712258 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svqnq\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-kube-api-access-svqnq\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.713691 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-etc-swift\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.722069 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-config-data\") pod \"swift-proxy-689b7bfc65-h98mj\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.799682 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.800190 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="ceilometer-central-agent" containerID="cri-o://6c539e3b0541d5b88f8cfaddae50496a0032906b15d6865b7f041b3968483932" gracePeriod=30 Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.800252 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="proxy-httpd" containerID="cri-o://1923fadcaf39e74964ba821ad8477e6764522a15f42da72ff27733773c8c2ccb" gracePeriod=30 Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.800277 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="sg-core" containerID="cri-o://53f1471876d0c8657cd1db9196437c738bda3153eb6235f76362ca2f52d69714" gracePeriod=30 Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.800308 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="ceilometer-notification-agent" containerID="cri-o://dd4d3eb043ce38a84be808f85db30333b93ef2b88d5dcd4c6bc3e1e69fe6661f" gracePeriod=30 Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.814715 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 28 11:44:39 crc kubenswrapper[4862]: I1128 11:44:39.838155 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:40 crc kubenswrapper[4862]: I1128 11:44:40.420545 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-689b7bfc65-h98mj"] Nov 28 11:44:40 crc kubenswrapper[4862]: W1128 11:44:40.434791 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedf3f6fa_2e74_489d_a32e_9faa5c0991ac.slice/crio-c9ab625d86195b4566d1e54b72e2ef06cfb59eabd688fffe663d285b8371c8fc WatchSource:0}: Error finding container c9ab625d86195b4566d1e54b72e2ef06cfb59eabd688fffe663d285b8371c8fc: Status 404 returned error can't find the container with id c9ab625d86195b4566d1e54b72e2ef06cfb59eabd688fffe663d285b8371c8fc Nov 28 11:44:40 crc kubenswrapper[4862]: I1128 11:44:40.779201 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-689b7bfc65-h98mj" event={"ID":"edf3f6fa-2e74-489d-a32e-9faa5c0991ac","Type":"ContainerStarted","Data":"b66e34c72b94210fa8e1b94676bc9d86fdb52829035fa8a84ba5bbf0ebf630c5"} Nov 28 11:44:40 crc kubenswrapper[4862]: I1128 11:44:40.779526 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-689b7bfc65-h98mj" event={"ID":"edf3f6fa-2e74-489d-a32e-9faa5c0991ac","Type":"ContainerStarted","Data":"c9ab625d86195b4566d1e54b72e2ef06cfb59eabd688fffe663d285b8371c8fc"} Nov 28 11:44:40 crc kubenswrapper[4862]: I1128 11:44:40.788491 4862 generic.go:334] "Generic (PLEG): container finished" podID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerID="1923fadcaf39e74964ba821ad8477e6764522a15f42da72ff27733773c8c2ccb" exitCode=0 Nov 28 11:44:40 crc kubenswrapper[4862]: I1128 11:44:40.788531 4862 generic.go:334] "Generic (PLEG): container finished" podID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerID="53f1471876d0c8657cd1db9196437c738bda3153eb6235f76362ca2f52d69714" exitCode=2 Nov 28 11:44:40 crc kubenswrapper[4862]: I1128 11:44:40.788540 4862 generic.go:334] "Generic (PLEG): container finished" podID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerID="6c539e3b0541d5b88f8cfaddae50496a0032906b15d6865b7f041b3968483932" exitCode=0 Nov 28 11:44:40 crc kubenswrapper[4862]: I1128 11:44:40.788526 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerDied","Data":"1923fadcaf39e74964ba821ad8477e6764522a15f42da72ff27733773c8c2ccb"} Nov 28 11:44:40 crc kubenswrapper[4862]: I1128 11:44:40.788595 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerDied","Data":"53f1471876d0c8657cd1db9196437c738bda3153eb6235f76362ca2f52d69714"} Nov 28 11:44:40 crc kubenswrapper[4862]: I1128 11:44:40.788615 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerDied","Data":"6c539e3b0541d5b88f8cfaddae50496a0032906b15d6865b7f041b3968483932"} Nov 28 11:44:41 crc kubenswrapper[4862]: I1128 11:44:41.805749 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-689b7bfc65-h98mj" event={"ID":"edf3f6fa-2e74-489d-a32e-9faa5c0991ac","Type":"ContainerStarted","Data":"cb861c743ce00570c223afefb40f3cc5ed244b9682a5c2e0bbf669dba92ca34f"} Nov 28 11:44:41 crc kubenswrapper[4862]: I1128 11:44:41.806056 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:41 crc kubenswrapper[4862]: I1128 11:44:41.844164 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-689b7bfc65-h98mj" podStartSLOduration=2.8441407290000003 podStartE2EDuration="2.844140729s" podCreationTimestamp="2025-11-28 11:44:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:41.836927611 +0000 UTC m=+1359.369441522" watchObservedRunningTime="2025-11-28 11:44:41.844140729 +0000 UTC m=+1359.376654650" Nov 28 11:44:42 crc kubenswrapper[4862]: I1128 11:44:42.813914 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:43 crc kubenswrapper[4862]: I1128 11:44:43.824912 4862 generic.go:334] "Generic (PLEG): container finished" podID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerID="dd4d3eb043ce38a84be808f85db30333b93ef2b88d5dcd4c6bc3e1e69fe6661f" exitCode=0 Nov 28 11:44:43 crc kubenswrapper[4862]: I1128 11:44:43.825069 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerDied","Data":"dd4d3eb043ce38a84be808f85db30333b93ef2b88d5dcd4c6bc3e1e69fe6661f"} Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.634772 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5jd4s"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.637280 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.650192 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5jd4s"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.731337 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-phtm9"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.732968 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.743686 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-144c-account-create-update-s9p9x"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.744943 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.746386 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9zd6\" (UniqueName: \"kubernetes.io/projected/88411d96-8709-4e1a-9042-a84a33008a18-kube-api-access-s9zd6\") pod \"nova-api-db-create-5jd4s\" (UID: \"88411d96-8709-4e1a-9042-a84a33008a18\") " pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.746550 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88411d96-8709-4e1a-9042-a84a33008a18-operator-scripts\") pod \"nova-api-db-create-5jd4s\" (UID: \"88411d96-8709-4e1a-9042-a84a33008a18\") " pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.748631 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.753563 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-phtm9"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.770359 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-144c-account-create-update-s9p9x"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.834863 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-srt25"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.841145 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.848563 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9zd6\" (UniqueName: \"kubernetes.io/projected/88411d96-8709-4e1a-9042-a84a33008a18-kube-api-access-s9zd6\") pod \"nova-api-db-create-5jd4s\" (UID: \"88411d96-8709-4e1a-9042-a84a33008a18\") " pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.848613 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3772cd33-8cc8-47a2-8453-f2395f011239-operator-scripts\") pod \"nova-api-144c-account-create-update-s9p9x\" (UID: \"3772cd33-8cc8-47a2-8453-f2395f011239\") " pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.848653 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88411d96-8709-4e1a-9042-a84a33008a18-operator-scripts\") pod \"nova-api-db-create-5jd4s\" (UID: \"88411d96-8709-4e1a-9042-a84a33008a18\") " pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.848683 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e1bd20-ac01-449f-b10c-aeba36373db9-operator-scripts\") pod \"nova-cell0-db-create-phtm9\" (UID: \"a1e1bd20-ac01-449f-b10c-aeba36373db9\") " pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.848707 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv5nd\" (UniqueName: \"kubernetes.io/projected/a1e1bd20-ac01-449f-b10c-aeba36373db9-kube-api-access-fv5nd\") pod \"nova-cell0-db-create-phtm9\" (UID: \"a1e1bd20-ac01-449f-b10c-aeba36373db9\") " pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.848723 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvqg5\" (UniqueName: \"kubernetes.io/projected/3772cd33-8cc8-47a2-8453-f2395f011239-kube-api-access-jvqg5\") pod \"nova-api-144c-account-create-update-s9p9x\" (UID: \"3772cd33-8cc8-47a2-8453-f2395f011239\") " pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.849643 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88411d96-8709-4e1a-9042-a84a33008a18-operator-scripts\") pod \"nova-api-db-create-5jd4s\" (UID: \"88411d96-8709-4e1a-9042-a84a33008a18\") " pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.856451 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-srt25"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.878121 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9zd6\" (UniqueName: \"kubernetes.io/projected/88411d96-8709-4e1a-9042-a84a33008a18-kube-api-access-s9zd6\") pod \"nova-api-db-create-5jd4s\" (UID: \"88411d96-8709-4e1a-9042-a84a33008a18\") " pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.942232 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-c5d7-account-create-update-8bflf"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.943294 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.952136 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-c5d7-account-create-update-8bflf"] Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.955891 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.953106 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv5nd\" (UniqueName: \"kubernetes.io/projected/a1e1bd20-ac01-449f-b10c-aeba36373db9-kube-api-access-fv5nd\") pod \"nova-cell0-db-create-phtm9\" (UID: \"a1e1bd20-ac01-449f-b10c-aeba36373db9\") " pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.956147 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvqg5\" (UniqueName: \"kubernetes.io/projected/3772cd33-8cc8-47a2-8453-f2395f011239-kube-api-access-jvqg5\") pod \"nova-api-144c-account-create-update-s9p9x\" (UID: \"3772cd33-8cc8-47a2-8453-f2395f011239\") " pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.956246 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06da04f0-f10f-4347-834a-88655ffab6fe-operator-scripts\") pod \"nova-cell1-db-create-srt25\" (UID: \"06da04f0-f10f-4347-834a-88655ffab6fe\") " pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.952812 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.956587 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95dc7\" (UniqueName: \"kubernetes.io/projected/06da04f0-f10f-4347-834a-88655ffab6fe-kube-api-access-95dc7\") pod \"nova-cell1-db-create-srt25\" (UID: \"06da04f0-f10f-4347-834a-88655ffab6fe\") " pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.956637 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3772cd33-8cc8-47a2-8453-f2395f011239-operator-scripts\") pod \"nova-api-144c-account-create-update-s9p9x\" (UID: \"3772cd33-8cc8-47a2-8453-f2395f011239\") " pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.956794 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e1bd20-ac01-449f-b10c-aeba36373db9-operator-scripts\") pod \"nova-cell0-db-create-phtm9\" (UID: \"a1e1bd20-ac01-449f-b10c-aeba36373db9\") " pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.957299 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3772cd33-8cc8-47a2-8453-f2395f011239-operator-scripts\") pod \"nova-api-144c-account-create-update-s9p9x\" (UID: \"3772cd33-8cc8-47a2-8453-f2395f011239\") " pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.957502 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e1bd20-ac01-449f-b10c-aeba36373db9-operator-scripts\") pod \"nova-cell0-db-create-phtm9\" (UID: \"a1e1bd20-ac01-449f-b10c-aeba36373db9\") " pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.975198 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvqg5\" (UniqueName: \"kubernetes.io/projected/3772cd33-8cc8-47a2-8453-f2395f011239-kube-api-access-jvqg5\") pod \"nova-api-144c-account-create-update-s9p9x\" (UID: \"3772cd33-8cc8-47a2-8453-f2395f011239\") " pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:47 crc kubenswrapper[4862]: I1128 11:44:47.978361 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv5nd\" (UniqueName: \"kubernetes.io/projected/a1e1bd20-ac01-449f-b10c-aeba36373db9-kube-api-access-fv5nd\") pod \"nova-cell0-db-create-phtm9\" (UID: \"a1e1bd20-ac01-449f-b10c-aeba36373db9\") " pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.052303 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.058210 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06da04f0-f10f-4347-834a-88655ffab6fe-operator-scripts\") pod \"nova-cell1-db-create-srt25\" (UID: \"06da04f0-f10f-4347-834a-88655ffab6fe\") " pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.058304 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de88799-efb7-4f0e-8c25-ace191d5885e-operator-scripts\") pod \"nova-cell0-c5d7-account-create-update-8bflf\" (UID: \"1de88799-efb7-4f0e-8c25-ace191d5885e\") " pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.058347 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lrwv\" (UniqueName: \"kubernetes.io/projected/1de88799-efb7-4f0e-8c25-ace191d5885e-kube-api-access-9lrwv\") pod \"nova-cell0-c5d7-account-create-update-8bflf\" (UID: \"1de88799-efb7-4f0e-8c25-ace191d5885e\") " pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.058477 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95dc7\" (UniqueName: \"kubernetes.io/projected/06da04f0-f10f-4347-834a-88655ffab6fe-kube-api-access-95dc7\") pod \"nova-cell1-db-create-srt25\" (UID: \"06da04f0-f10f-4347-834a-88655ffab6fe\") " pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.059500 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06da04f0-f10f-4347-834a-88655ffab6fe-operator-scripts\") pod \"nova-cell1-db-create-srt25\" (UID: \"06da04f0-f10f-4347-834a-88655ffab6fe\") " pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.069717 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.074799 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95dc7\" (UniqueName: \"kubernetes.io/projected/06da04f0-f10f-4347-834a-88655ffab6fe-kube-api-access-95dc7\") pod \"nova-cell1-db-create-srt25\" (UID: \"06da04f0-f10f-4347-834a-88655ffab6fe\") " pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.154623 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.158627 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-c990-account-create-update-2sxzh"] Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.160263 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.160279 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de88799-efb7-4f0e-8c25-ace191d5885e-operator-scripts\") pod \"nova-cell0-c5d7-account-create-update-8bflf\" (UID: \"1de88799-efb7-4f0e-8c25-ace191d5885e\") " pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.160338 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lrwv\" (UniqueName: \"kubernetes.io/projected/1de88799-efb7-4f0e-8c25-ace191d5885e-kube-api-access-9lrwv\") pod \"nova-cell0-c5d7-account-create-update-8bflf\" (UID: \"1de88799-efb7-4f0e-8c25-ace191d5885e\") " pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.161109 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de88799-efb7-4f0e-8c25-ace191d5885e-operator-scripts\") pod \"nova-cell0-c5d7-account-create-update-8bflf\" (UID: \"1de88799-efb7-4f0e-8c25-ace191d5885e\") " pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.170159 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.171103 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c990-account-create-update-2sxzh"] Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.184573 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lrwv\" (UniqueName: \"kubernetes.io/projected/1de88799-efb7-4f0e-8c25-ace191d5885e-kube-api-access-9lrwv\") pod \"nova-cell0-c5d7-account-create-update-8bflf\" (UID: \"1de88799-efb7-4f0e-8c25-ace191d5885e\") " pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.262767 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7x5z\" (UniqueName: \"kubernetes.io/projected/30442d53-db04-4396-b26a-0605fa91c58f-kube-api-access-q7x5z\") pod \"nova-cell1-c990-account-create-update-2sxzh\" (UID: \"30442d53-db04-4396-b26a-0605fa91c58f\") " pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.262938 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30442d53-db04-4396-b26a-0605fa91c58f-operator-scripts\") pod \"nova-cell1-c990-account-create-update-2sxzh\" (UID: \"30442d53-db04-4396-b26a-0605fa91c58f\") " pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.265575 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.365945 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30442d53-db04-4396-b26a-0605fa91c58f-operator-scripts\") pod \"nova-cell1-c990-account-create-update-2sxzh\" (UID: \"30442d53-db04-4396-b26a-0605fa91c58f\") " pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.366025 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7x5z\" (UniqueName: \"kubernetes.io/projected/30442d53-db04-4396-b26a-0605fa91c58f-kube-api-access-q7x5z\") pod \"nova-cell1-c990-account-create-update-2sxzh\" (UID: \"30442d53-db04-4396-b26a-0605fa91c58f\") " pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.366852 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30442d53-db04-4396-b26a-0605fa91c58f-operator-scripts\") pod \"nova-cell1-c990-account-create-update-2sxzh\" (UID: \"30442d53-db04-4396-b26a-0605fa91c58f\") " pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.382203 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7x5z\" (UniqueName: \"kubernetes.io/projected/30442d53-db04-4396-b26a-0605fa91c58f-kube-api-access-q7x5z\") pod \"nova-cell1-c990-account-create-update-2sxzh\" (UID: \"30442d53-db04-4396-b26a-0605fa91c58f\") " pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:48 crc kubenswrapper[4862]: I1128 11:44:48.530341 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:49 crc kubenswrapper[4862]: I1128 11:44:49.844371 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:49 crc kubenswrapper[4862]: I1128 11:44:49.847913 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.131747 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.217558 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fwz9\" (UniqueName: \"kubernetes.io/projected/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-kube-api-access-8fwz9\") pod \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.217899 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-run-httpd\") pod \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.217967 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-combined-ca-bundle\") pod \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.217996 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-sg-core-conf-yaml\") pod \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.218026 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-scripts\") pod \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.218061 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-config-data\") pod \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.218081 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-log-httpd\") pod \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\" (UID: \"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a\") " Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.219278 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" (UID: "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.219531 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" (UID: "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.223779 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-kube-api-access-8fwz9" (OuterVolumeSpecName: "kube-api-access-8fwz9") pod "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" (UID: "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a"). InnerVolumeSpecName "kube-api-access-8fwz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.236246 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-scripts" (OuterVolumeSpecName: "scripts") pod "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" (UID: "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.251983 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" (UID: "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.324876 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fwz9\" (UniqueName: \"kubernetes.io/projected/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-kube-api-access-8fwz9\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.324909 4862 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.324920 4862 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.324931 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.324942 4862 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.335325 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" (UID: "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.387736 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-config-data" (OuterVolumeSpecName: "config-data") pod "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" (UID: "bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.426721 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.426762 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.505265 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-144c-account-create-update-s9p9x"] Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.872840 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5jd4s"] Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.886731 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-srt25"] Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.894636 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-c5d7-account-create-update-8bflf"] Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.925737 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"23891b0e-3fc7-4d2f-9abf-96b57b09ca01","Type":"ContainerStarted","Data":"36eb5af2c95a5dd49963c18b778f72ccbf03d09076fbe15508a1df0b8532d372"} Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.930858 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5jd4s" event={"ID":"88411d96-8709-4e1a-9042-a84a33008a18","Type":"ContainerStarted","Data":"2b0842afd209da1552ca7edd1e8650bafd8b98884e0a59406abe940a05c368a8"} Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.934339 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-srt25" event={"ID":"06da04f0-f10f-4347-834a-88655ffab6fe","Type":"ContainerStarted","Data":"0013d3f2102072d040c54f4719d6e29f39ae884421dc0a7341be66143c69432b"} Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.936973 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-144c-account-create-update-s9p9x" event={"ID":"3772cd33-8cc8-47a2-8453-f2395f011239","Type":"ContainerStarted","Data":"44535d7cc32f4331f5fc1f21419dfd1a2f555c351ab9752b1439e4c14dca91c9"} Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.937059 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-144c-account-create-update-s9p9x" event={"ID":"3772cd33-8cc8-47a2-8453-f2395f011239","Type":"ContainerStarted","Data":"c010542bfada391c5efd2b9c212c1694ea40a7eed1371068429720d9f31e9af0"} Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.955788 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.852107854 podStartE2EDuration="15.955762946s" podCreationTimestamp="2025-11-28 11:44:36 +0000 UTC" firstStartedPulling="2025-11-28 11:44:37.061957229 +0000 UTC m=+1354.594471150" lastFinishedPulling="2025-11-28 11:44:51.165612321 +0000 UTC m=+1368.698126242" observedRunningTime="2025-11-28 11:44:51.940869936 +0000 UTC m=+1369.473383867" watchObservedRunningTime="2025-11-28 11:44:51.955762946 +0000 UTC m=+1369.488276877" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.962583 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a","Type":"ContainerDied","Data":"1c97dfaf0d885762cecbf7531e0bdc9dc44a22faa74b1ef5ba0884a8cf0e8117"} Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.962651 4862 scope.go:117] "RemoveContainer" containerID="1923fadcaf39e74964ba821ad8477e6764522a15f42da72ff27733773c8c2ccb" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.962845 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.966820 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" event={"ID":"1de88799-efb7-4f0e-8c25-ace191d5885e","Type":"ContainerStarted","Data":"04c04a05144a021c4486e4ef7ff496e335f169bdd4494b7472d00b27bc9111b8"} Nov 28 11:44:51 crc kubenswrapper[4862]: I1128 11:44:51.970380 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-144c-account-create-update-s9p9x" podStartSLOduration=4.970359097 podStartE2EDuration="4.970359097s" podCreationTimestamp="2025-11-28 11:44:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:44:51.957911018 +0000 UTC m=+1369.490424939" watchObservedRunningTime="2025-11-28 11:44:51.970359097 +0000 UTC m=+1369.502873018" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.013952 4862 scope.go:117] "RemoveContainer" containerID="53f1471876d0c8657cd1db9196437c738bda3153eb6235f76362ca2f52d69714" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.040946 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c990-account-create-update-2sxzh"] Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.059326 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-phtm9"] Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.295647 4862 scope.go:117] "RemoveContainer" containerID="dd4d3eb043ce38a84be808f85db30333b93ef2b88d5dcd4c6bc3e1e69fe6661f" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.311199 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.325712 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.335243 4862 scope.go:117] "RemoveContainer" containerID="6c539e3b0541d5b88f8cfaddae50496a0032906b15d6865b7f041b3968483932" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.363729 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:52 crc kubenswrapper[4862]: E1128 11:44:52.364245 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="ceilometer-central-agent" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.364262 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="ceilometer-central-agent" Nov 28 11:44:52 crc kubenswrapper[4862]: E1128 11:44:52.364279 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="proxy-httpd" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.364285 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="proxy-httpd" Nov 28 11:44:52 crc kubenswrapper[4862]: E1128 11:44:52.364301 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="sg-core" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.364306 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="sg-core" Nov 28 11:44:52 crc kubenswrapper[4862]: E1128 11:44:52.364327 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="ceilometer-notification-agent" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.364333 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="ceilometer-notification-agent" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.364610 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="sg-core" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.364638 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="ceilometer-notification-agent" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.364651 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="ceilometer-central-agent" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.364732 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="proxy-httpd" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.366356 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.374487 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.374742 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.405481 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.551866 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.551904 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-log-httpd\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.551935 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdtcz\" (UniqueName: \"kubernetes.io/projected/b26fee90-910d-4715-bdbd-4fe6d1695d4c-kube-api-access-jdtcz\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.552166 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.552268 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-run-httpd\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.552480 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-scripts\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.552557 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-config-data\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.653779 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.653843 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-run-httpd\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.653900 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-scripts\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.653917 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-config-data\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.653983 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.654001 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-log-httpd\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.654022 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdtcz\" (UniqueName: \"kubernetes.io/projected/b26fee90-910d-4715-bdbd-4fe6d1695d4c-kube-api-access-jdtcz\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.654381 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-run-httpd\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.660914 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.660939 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-config-data\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.661212 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-log-httpd\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.664743 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.665285 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-scripts\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.676587 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdtcz\" (UniqueName: \"kubernetes.io/projected/b26fee90-910d-4715-bdbd-4fe6d1695d4c-kube-api-access-jdtcz\") pod \"ceilometer-0\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.732980 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.863231 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" path="/var/lib/kubelet/pods/bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a/volumes" Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.986560 4862 generic.go:334] "Generic (PLEG): container finished" podID="a1e1bd20-ac01-449f-b10c-aeba36373db9" containerID="32319cdb56353b11cfea025f3670b75833c0097861bc5bbd3e48e49d8ea01757" exitCode=0 Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.986616 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-phtm9" event={"ID":"a1e1bd20-ac01-449f-b10c-aeba36373db9","Type":"ContainerDied","Data":"32319cdb56353b11cfea025f3670b75833c0097861bc5bbd3e48e49d8ea01757"} Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.986642 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-phtm9" event={"ID":"a1e1bd20-ac01-449f-b10c-aeba36373db9","Type":"ContainerStarted","Data":"c181ffec3aff17e9e1bba14b03e41689351399c670453e73e71d42d390d85bdb"} Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.989403 4862 generic.go:334] "Generic (PLEG): container finished" podID="88411d96-8709-4e1a-9042-a84a33008a18" containerID="846a82b096c91fa279b8a2c68e6b0f950e08a8bae9936e49194d3d9a2af99b3b" exitCode=0 Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.989486 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5jd4s" event={"ID":"88411d96-8709-4e1a-9042-a84a33008a18","Type":"ContainerDied","Data":"846a82b096c91fa279b8a2c68e6b0f950e08a8bae9936e49194d3d9a2af99b3b"} Nov 28 11:44:52 crc kubenswrapper[4862]: I1128 11:44:52.995990 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-srt25" event={"ID":"06da04f0-f10f-4347-834a-88655ffab6fe","Type":"ContainerDied","Data":"e2ae775b529fe1449b01a84b2dfacab55da288eac72f7eaca548c563190f06b2"} Nov 28 11:44:53 crc kubenswrapper[4862]: I1128 11:44:52.996068 4862 generic.go:334] "Generic (PLEG): container finished" podID="06da04f0-f10f-4347-834a-88655ffab6fe" containerID="e2ae775b529fe1449b01a84b2dfacab55da288eac72f7eaca548c563190f06b2" exitCode=0 Nov 28 11:44:53 crc kubenswrapper[4862]: I1128 11:44:53.004528 4862 generic.go:334] "Generic (PLEG): container finished" podID="3772cd33-8cc8-47a2-8453-f2395f011239" containerID="44535d7cc32f4331f5fc1f21419dfd1a2f555c351ab9752b1439e4c14dca91c9" exitCode=0 Nov 28 11:44:53 crc kubenswrapper[4862]: I1128 11:44:53.004612 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-144c-account-create-update-s9p9x" event={"ID":"3772cd33-8cc8-47a2-8453-f2395f011239","Type":"ContainerDied","Data":"44535d7cc32f4331f5fc1f21419dfd1a2f555c351ab9752b1439e4c14dca91c9"} Nov 28 11:44:53 crc kubenswrapper[4862]: I1128 11:44:53.009442 4862 generic.go:334] "Generic (PLEG): container finished" podID="1de88799-efb7-4f0e-8c25-ace191d5885e" containerID="16d4372c7aab7bdaa122d6237f66d4b150a71493de60bf9bed530c808ad0b1a1" exitCode=0 Nov 28 11:44:53 crc kubenswrapper[4862]: I1128 11:44:53.009518 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" event={"ID":"1de88799-efb7-4f0e-8c25-ace191d5885e","Type":"ContainerDied","Data":"16d4372c7aab7bdaa122d6237f66d4b150a71493de60bf9bed530c808ad0b1a1"} Nov 28 11:44:53 crc kubenswrapper[4862]: I1128 11:44:53.014874 4862 generic.go:334] "Generic (PLEG): container finished" podID="30442d53-db04-4396-b26a-0605fa91c58f" containerID="ca4211ba57f4a2e251d0702c6e45025532c02da611b321564e0087fbd70a912b" exitCode=0 Nov 28 11:44:53 crc kubenswrapper[4862]: I1128 11:44:53.017357 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c990-account-create-update-2sxzh" event={"ID":"30442d53-db04-4396-b26a-0605fa91c58f","Type":"ContainerDied","Data":"ca4211ba57f4a2e251d0702c6e45025532c02da611b321564e0087fbd70a912b"} Nov 28 11:44:53 crc kubenswrapper[4862]: I1128 11:44:53.021210 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c990-account-create-update-2sxzh" event={"ID":"30442d53-db04-4396-b26a-0605fa91c58f","Type":"ContainerStarted","Data":"94b76c4fa3030601b16ff6f66822dfbe9f7829e71e88f9f849f9570833231dba"} Nov 28 11:44:53 crc kubenswrapper[4862]: I1128 11:44:53.126860 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.052749 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerStarted","Data":"3bd7367ad3c8efcebb190a96352d52553a2cf7932124cdf7f8aa03f69869583a"} Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.053686 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerStarted","Data":"e52e0e1ae45980704418392b04dde3c69fcb570f0a0eef0e6d0bdad0da765d3b"} Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.590818 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.668773 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.676799 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.705220 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06da04f0-f10f-4347-834a-88655ffab6fe-operator-scripts\") pod \"06da04f0-f10f-4347-834a-88655ffab6fe\" (UID: \"06da04f0-f10f-4347-834a-88655ffab6fe\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.705434 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95dc7\" (UniqueName: \"kubernetes.io/projected/06da04f0-f10f-4347-834a-88655ffab6fe-kube-api-access-95dc7\") pod \"06da04f0-f10f-4347-834a-88655ffab6fe\" (UID: \"06da04f0-f10f-4347-834a-88655ffab6fe\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.706537 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06da04f0-f10f-4347-834a-88655ffab6fe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "06da04f0-f10f-4347-834a-88655ffab6fe" (UID: "06da04f0-f10f-4347-834a-88655ffab6fe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.710963 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06da04f0-f10f-4347-834a-88655ffab6fe-kube-api-access-95dc7" (OuterVolumeSpecName: "kube-api-access-95dc7") pod "06da04f0-f10f-4347-834a-88655ffab6fe" (UID: "06da04f0-f10f-4347-834a-88655ffab6fe"). InnerVolumeSpecName "kube-api-access-95dc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.716951 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.736478 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.752570 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.806774 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lrwv\" (UniqueName: \"kubernetes.io/projected/1de88799-efb7-4f0e-8c25-ace191d5885e-kube-api-access-9lrwv\") pod \"1de88799-efb7-4f0e-8c25-ace191d5885e\" (UID: \"1de88799-efb7-4f0e-8c25-ace191d5885e\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.806902 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de88799-efb7-4f0e-8c25-ace191d5885e-operator-scripts\") pod \"1de88799-efb7-4f0e-8c25-ace191d5885e\" (UID: \"1de88799-efb7-4f0e-8c25-ace191d5885e\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.807058 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30442d53-db04-4396-b26a-0605fa91c58f-operator-scripts\") pod \"30442d53-db04-4396-b26a-0605fa91c58f\" (UID: \"30442d53-db04-4396-b26a-0605fa91c58f\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.807082 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7x5z\" (UniqueName: \"kubernetes.io/projected/30442d53-db04-4396-b26a-0605fa91c58f-kube-api-access-q7x5z\") pod \"30442d53-db04-4396-b26a-0605fa91c58f\" (UID: \"30442d53-db04-4396-b26a-0605fa91c58f\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.807515 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95dc7\" (UniqueName: \"kubernetes.io/projected/06da04f0-f10f-4347-834a-88655ffab6fe-kube-api-access-95dc7\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.807527 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06da04f0-f10f-4347-834a-88655ffab6fe-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.807590 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30442d53-db04-4396-b26a-0605fa91c58f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "30442d53-db04-4396-b26a-0605fa91c58f" (UID: "30442d53-db04-4396-b26a-0605fa91c58f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.807721 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de88799-efb7-4f0e-8c25-ace191d5885e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1de88799-efb7-4f0e-8c25-ace191d5885e" (UID: "1de88799-efb7-4f0e-8c25-ace191d5885e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.811414 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de88799-efb7-4f0e-8c25-ace191d5885e-kube-api-access-9lrwv" (OuterVolumeSpecName: "kube-api-access-9lrwv") pod "1de88799-efb7-4f0e-8c25-ace191d5885e" (UID: "1de88799-efb7-4f0e-8c25-ace191d5885e"). InnerVolumeSpecName "kube-api-access-9lrwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.812522 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30442d53-db04-4396-b26a-0605fa91c58f-kube-api-access-q7x5z" (OuterVolumeSpecName: "kube-api-access-q7x5z") pod "30442d53-db04-4396-b26a-0605fa91c58f" (UID: "30442d53-db04-4396-b26a-0605fa91c58f"). InnerVolumeSpecName "kube-api-access-q7x5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.909013 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88411d96-8709-4e1a-9042-a84a33008a18-operator-scripts\") pod \"88411d96-8709-4e1a-9042-a84a33008a18\" (UID: \"88411d96-8709-4e1a-9042-a84a33008a18\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.909109 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e1bd20-ac01-449f-b10c-aeba36373db9-operator-scripts\") pod \"a1e1bd20-ac01-449f-b10c-aeba36373db9\" (UID: \"a1e1bd20-ac01-449f-b10c-aeba36373db9\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.909185 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv5nd\" (UniqueName: \"kubernetes.io/projected/a1e1bd20-ac01-449f-b10c-aeba36373db9-kube-api-access-fv5nd\") pod \"a1e1bd20-ac01-449f-b10c-aeba36373db9\" (UID: \"a1e1bd20-ac01-449f-b10c-aeba36373db9\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.909240 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvqg5\" (UniqueName: \"kubernetes.io/projected/3772cd33-8cc8-47a2-8453-f2395f011239-kube-api-access-jvqg5\") pod \"3772cd33-8cc8-47a2-8453-f2395f011239\" (UID: \"3772cd33-8cc8-47a2-8453-f2395f011239\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.909315 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3772cd33-8cc8-47a2-8453-f2395f011239-operator-scripts\") pod \"3772cd33-8cc8-47a2-8453-f2395f011239\" (UID: \"3772cd33-8cc8-47a2-8453-f2395f011239\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.909419 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9zd6\" (UniqueName: \"kubernetes.io/projected/88411d96-8709-4e1a-9042-a84a33008a18-kube-api-access-s9zd6\") pod \"88411d96-8709-4e1a-9042-a84a33008a18\" (UID: \"88411d96-8709-4e1a-9042-a84a33008a18\") " Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.910054 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lrwv\" (UniqueName: \"kubernetes.io/projected/1de88799-efb7-4f0e-8c25-ace191d5885e-kube-api-access-9lrwv\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.910081 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de88799-efb7-4f0e-8c25-ace191d5885e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.910112 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30442d53-db04-4396-b26a-0605fa91c58f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.910127 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7x5z\" (UniqueName: \"kubernetes.io/projected/30442d53-db04-4396-b26a-0605fa91c58f-kube-api-access-q7x5z\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.913450 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88411d96-8709-4e1a-9042-a84a33008a18-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "88411d96-8709-4e1a-9042-a84a33008a18" (UID: "88411d96-8709-4e1a-9042-a84a33008a18"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.914018 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88411d96-8709-4e1a-9042-a84a33008a18-kube-api-access-s9zd6" (OuterVolumeSpecName: "kube-api-access-s9zd6") pod "88411d96-8709-4e1a-9042-a84a33008a18" (UID: "88411d96-8709-4e1a-9042-a84a33008a18"). InnerVolumeSpecName "kube-api-access-s9zd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.914568 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3772cd33-8cc8-47a2-8453-f2395f011239-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3772cd33-8cc8-47a2-8453-f2395f011239" (UID: "3772cd33-8cc8-47a2-8453-f2395f011239"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.915012 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3772cd33-8cc8-47a2-8453-f2395f011239-kube-api-access-jvqg5" (OuterVolumeSpecName: "kube-api-access-jvqg5") pod "3772cd33-8cc8-47a2-8453-f2395f011239" (UID: "3772cd33-8cc8-47a2-8453-f2395f011239"). InnerVolumeSpecName "kube-api-access-jvqg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.917954 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1e1bd20-ac01-449f-b10c-aeba36373db9-kube-api-access-fv5nd" (OuterVolumeSpecName: "kube-api-access-fv5nd") pod "a1e1bd20-ac01-449f-b10c-aeba36373db9" (UID: "a1e1bd20-ac01-449f-b10c-aeba36373db9"). InnerVolumeSpecName "kube-api-access-fv5nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:44:54 crc kubenswrapper[4862]: I1128 11:44:54.918256 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1e1bd20-ac01-449f-b10c-aeba36373db9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a1e1bd20-ac01-449f-b10c-aeba36373db9" (UID: "a1e1bd20-ac01-449f-b10c-aeba36373db9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.011969 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3772cd33-8cc8-47a2-8453-f2395f011239-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.012017 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9zd6\" (UniqueName: \"kubernetes.io/projected/88411d96-8709-4e1a-9042-a84a33008a18-kube-api-access-s9zd6\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.012032 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88411d96-8709-4e1a-9042-a84a33008a18-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.012043 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e1bd20-ac01-449f-b10c-aeba36373db9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.012055 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv5nd\" (UniqueName: \"kubernetes.io/projected/a1e1bd20-ac01-449f-b10c-aeba36373db9-kube-api-access-fv5nd\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.012068 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvqg5\" (UniqueName: \"kubernetes.io/projected/3772cd33-8cc8-47a2-8453-f2395f011239-kube-api-access-jvqg5\") on node \"crc\" DevicePath \"\"" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.071219 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5jd4s" event={"ID":"88411d96-8709-4e1a-9042-a84a33008a18","Type":"ContainerDied","Data":"2b0842afd209da1552ca7edd1e8650bafd8b98884e0a59406abe940a05c368a8"} Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.071253 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5jd4s" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.071257 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b0842afd209da1552ca7edd1e8650bafd8b98884e0a59406abe940a05c368a8" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.084289 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-srt25" event={"ID":"06da04f0-f10f-4347-834a-88655ffab6fe","Type":"ContainerDied","Data":"0013d3f2102072d040c54f4719d6e29f39ae884421dc0a7341be66143c69432b"} Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.084328 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0013d3f2102072d040c54f4719d6e29f39ae884421dc0a7341be66143c69432b" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.084331 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-srt25" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.087056 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-144c-account-create-update-s9p9x" event={"ID":"3772cd33-8cc8-47a2-8453-f2395f011239","Type":"ContainerDied","Data":"c010542bfada391c5efd2b9c212c1694ea40a7eed1371068429720d9f31e9af0"} Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.087124 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c010542bfada391c5efd2b9c212c1694ea40a7eed1371068429720d9f31e9af0" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.087172 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-144c-account-create-update-s9p9x" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.090771 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerStarted","Data":"f990be225595ff76feff0532ec85ced4559af58db51b56bec6908ff5195920c0"} Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.092436 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" event={"ID":"1de88799-efb7-4f0e-8c25-ace191d5885e","Type":"ContainerDied","Data":"04c04a05144a021c4486e4ef7ff496e335f169bdd4494b7472d00b27bc9111b8"} Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.092457 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04c04a05144a021c4486e4ef7ff496e335f169bdd4494b7472d00b27bc9111b8" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.092471 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-c5d7-account-create-update-8bflf" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.096932 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-phtm9" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.096952 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-phtm9" event={"ID":"a1e1bd20-ac01-449f-b10c-aeba36373db9","Type":"ContainerDied","Data":"c181ffec3aff17e9e1bba14b03e41689351399c670453e73e71d42d390d85bdb"} Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.096996 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c181ffec3aff17e9e1bba14b03e41689351399c670453e73e71d42d390d85bdb" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.104236 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c990-account-create-update-2sxzh" event={"ID":"30442d53-db04-4396-b26a-0605fa91c58f","Type":"ContainerDied","Data":"94b76c4fa3030601b16ff6f66822dfbe9f7829e71e88f9f849f9570833231dba"} Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.104293 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94b76c4fa3030601b16ff6f66822dfbe9f7829e71e88f9f849f9570833231dba" Nov 28 11:44:55 crc kubenswrapper[4862]: I1128 11:44:55.104353 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c990-account-create-update-2sxzh" Nov 28 11:44:56 crc kubenswrapper[4862]: I1128 11:44:56.116108 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerStarted","Data":"dbb316624926bb234d8295ddb2799f85b282e85bd10b8e727612c8fe23f64b7e"} Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.138320 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerStarted","Data":"abd79235cd914e6150ae2d639121566a928bbb2d11a66e837754fa48b4877def"} Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.138845 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.170043 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.311898627 podStartE2EDuration="6.170024839s" podCreationTimestamp="2025-11-28 11:44:52 +0000 UTC" firstStartedPulling="2025-11-28 11:44:53.108300325 +0000 UTC m=+1370.640814256" lastFinishedPulling="2025-11-28 11:44:56.966426507 +0000 UTC m=+1374.498940468" observedRunningTime="2025-11-28 11:44:58.168641479 +0000 UTC m=+1375.701155410" watchObservedRunningTime="2025-11-28 11:44:58.170024839 +0000 UTC m=+1375.702538770" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.223458 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2tm6f"] Nov 28 11:44:58 crc kubenswrapper[4862]: E1128 11:44:58.224029 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88411d96-8709-4e1a-9042-a84a33008a18" containerName="mariadb-database-create" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.224054 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="88411d96-8709-4e1a-9042-a84a33008a18" containerName="mariadb-database-create" Nov 28 11:44:58 crc kubenswrapper[4862]: E1128 11:44:58.224081 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30442d53-db04-4396-b26a-0605fa91c58f" containerName="mariadb-account-create-update" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.224665 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="30442d53-db04-4396-b26a-0605fa91c58f" containerName="mariadb-account-create-update" Nov 28 11:44:58 crc kubenswrapper[4862]: E1128 11:44:58.224683 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3772cd33-8cc8-47a2-8453-f2395f011239" containerName="mariadb-account-create-update" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.224694 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="3772cd33-8cc8-47a2-8453-f2395f011239" containerName="mariadb-account-create-update" Nov 28 11:44:58 crc kubenswrapper[4862]: E1128 11:44:58.224717 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06da04f0-f10f-4347-834a-88655ffab6fe" containerName="mariadb-database-create" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.224728 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="06da04f0-f10f-4347-834a-88655ffab6fe" containerName="mariadb-database-create" Nov 28 11:44:58 crc kubenswrapper[4862]: E1128 11:44:58.224744 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de88799-efb7-4f0e-8c25-ace191d5885e" containerName="mariadb-account-create-update" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.224752 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de88799-efb7-4f0e-8c25-ace191d5885e" containerName="mariadb-account-create-update" Nov 28 11:44:58 crc kubenswrapper[4862]: E1128 11:44:58.224772 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e1bd20-ac01-449f-b10c-aeba36373db9" containerName="mariadb-database-create" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.224781 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e1bd20-ac01-449f-b10c-aeba36373db9" containerName="mariadb-database-create" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.225010 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="3772cd33-8cc8-47a2-8453-f2395f011239" containerName="mariadb-account-create-update" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.225051 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="30442d53-db04-4396-b26a-0605fa91c58f" containerName="mariadb-account-create-update" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.225068 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="88411d96-8709-4e1a-9042-a84a33008a18" containerName="mariadb-database-create" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.225085 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="06da04f0-f10f-4347-834a-88655ffab6fe" containerName="mariadb-database-create" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.225115 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e1bd20-ac01-449f-b10c-aeba36373db9" containerName="mariadb-database-create" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.225135 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de88799-efb7-4f0e-8c25-ace191d5885e" containerName="mariadb-account-create-update" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.225817 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.227910 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.228196 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.228971 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xdvln" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.242033 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2tm6f"] Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.374635 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-scripts\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.374740 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-config-data\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.374792 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdxnq\" (UniqueName: \"kubernetes.io/projected/0cfe27ee-17c6-481a-81c0-6b48211db18c-kube-api-access-gdxnq\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.374833 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.476230 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.476607 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-scripts\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.476669 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-config-data\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.476713 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdxnq\" (UniqueName: \"kubernetes.io/projected/0cfe27ee-17c6-481a-81c0-6b48211db18c-kube-api-access-gdxnq\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.480846 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-scripts\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.484788 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.484840 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-config-data\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.503164 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdxnq\" (UniqueName: \"kubernetes.io/projected/0cfe27ee-17c6-481a-81c0-6b48211db18c-kube-api-access-gdxnq\") pod \"nova-cell0-conductor-db-sync-2tm6f\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: I1128 11:44:58.545336 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:44:58 crc kubenswrapper[4862]: W1128 11:44:58.995510 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cfe27ee_17c6_481a_81c0_6b48211db18c.slice/crio-f1a410ebb2478606e731dbad7b8280bddf0d97fee98531caad8715e0356f0410 WatchSource:0}: Error finding container f1a410ebb2478606e731dbad7b8280bddf0d97fee98531caad8715e0356f0410: Status 404 returned error can't find the container with id f1a410ebb2478606e731dbad7b8280bddf0d97fee98531caad8715e0356f0410 Nov 28 11:44:59 crc kubenswrapper[4862]: I1128 11:44:59.000297 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2tm6f"] Nov 28 11:44:59 crc kubenswrapper[4862]: I1128 11:44:59.147937 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2tm6f" event={"ID":"0cfe27ee-17c6-481a-81c0-6b48211db18c","Type":"ContainerStarted","Data":"f1a410ebb2478606e731dbad7b8280bddf0d97fee98531caad8715e0356f0410"} Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.134608 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss"] Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.137691 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.141275 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.143303 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.167397 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss"] Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.313562 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99f601b7-1297-410b-b3e1-41bc2a1bb60b-secret-volume\") pod \"collect-profiles-29405505-wf5ss\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.313633 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99f601b7-1297-410b-b3e1-41bc2a1bb60b-config-volume\") pod \"collect-profiles-29405505-wf5ss\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.313682 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvjvx\" (UniqueName: \"kubernetes.io/projected/99f601b7-1297-410b-b3e1-41bc2a1bb60b-kube-api-access-lvjvx\") pod \"collect-profiles-29405505-wf5ss\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.414876 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvjvx\" (UniqueName: \"kubernetes.io/projected/99f601b7-1297-410b-b3e1-41bc2a1bb60b-kube-api-access-lvjvx\") pod \"collect-profiles-29405505-wf5ss\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.415044 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99f601b7-1297-410b-b3e1-41bc2a1bb60b-secret-volume\") pod \"collect-profiles-29405505-wf5ss\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.415138 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99f601b7-1297-410b-b3e1-41bc2a1bb60b-config-volume\") pod \"collect-profiles-29405505-wf5ss\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.416183 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99f601b7-1297-410b-b3e1-41bc2a1bb60b-config-volume\") pod \"collect-profiles-29405505-wf5ss\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.421719 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99f601b7-1297-410b-b3e1-41bc2a1bb60b-secret-volume\") pod \"collect-profiles-29405505-wf5ss\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.434792 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvjvx\" (UniqueName: \"kubernetes.io/projected/99f601b7-1297-410b-b3e1-41bc2a1bb60b-kube-api-access-lvjvx\") pod \"collect-profiles-29405505-wf5ss\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:00 crc kubenswrapper[4862]: I1128 11:45:00.475932 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:01 crc kubenswrapper[4862]: I1128 11:45:01.766818 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss"] Nov 28 11:45:01 crc kubenswrapper[4862]: W1128 11:45:01.790564 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99f601b7_1297_410b_b3e1_41bc2a1bb60b.slice/crio-97b6a34de1be80c3c3c0940b5678e22f51719982cc605a8f067eaefdae4431d1 WatchSource:0}: Error finding container 97b6a34de1be80c3c3c0940b5678e22f51719982cc605a8f067eaefdae4431d1: Status 404 returned error can't find the container with id 97b6a34de1be80c3c3c0940b5678e22f51719982cc605a8f067eaefdae4431d1 Nov 28 11:45:02 crc kubenswrapper[4862]: I1128 11:45:02.199997 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" event={"ID":"99f601b7-1297-410b-b3e1-41bc2a1bb60b","Type":"ContainerStarted","Data":"480c94fc03f9c01637740cad79a65eccf564307680f1de6618acd5fa765470c2"} Nov 28 11:45:02 crc kubenswrapper[4862]: I1128 11:45:02.200525 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" event={"ID":"99f601b7-1297-410b-b3e1-41bc2a1bb60b","Type":"ContainerStarted","Data":"97b6a34de1be80c3c3c0940b5678e22f51719982cc605a8f067eaefdae4431d1"} Nov 28 11:45:02 crc kubenswrapper[4862]: I1128 11:45:02.223198 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" podStartSLOduration=2.223182417 podStartE2EDuration="2.223182417s" podCreationTimestamp="2025-11-28 11:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:02.214427514 +0000 UTC m=+1379.746941435" watchObservedRunningTime="2025-11-28 11:45:02.223182417 +0000 UTC m=+1379.755696338" Nov 28 11:45:03 crc kubenswrapper[4862]: I1128 11:45:03.233882 4862 generic.go:334] "Generic (PLEG): container finished" podID="99f601b7-1297-410b-b3e1-41bc2a1bb60b" containerID="480c94fc03f9c01637740cad79a65eccf564307680f1de6618acd5fa765470c2" exitCode=0 Nov 28 11:45:03 crc kubenswrapper[4862]: I1128 11:45:03.234183 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" event={"ID":"99f601b7-1297-410b-b3e1-41bc2a1bb60b","Type":"ContainerDied","Data":"480c94fc03f9c01637740cad79a65eccf564307680f1de6618acd5fa765470c2"} Nov 28 11:45:04 crc kubenswrapper[4862]: I1128 11:45:04.456264 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:04 crc kubenswrapper[4862]: I1128 11:45:04.456872 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="ceilometer-central-agent" containerID="cri-o://3bd7367ad3c8efcebb190a96352d52553a2cf7932124cdf7f8aa03f69869583a" gracePeriod=30 Nov 28 11:45:04 crc kubenswrapper[4862]: I1128 11:45:04.456991 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="proxy-httpd" containerID="cri-o://abd79235cd914e6150ae2d639121566a928bbb2d11a66e837754fa48b4877def" gracePeriod=30 Nov 28 11:45:04 crc kubenswrapper[4862]: I1128 11:45:04.457023 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="sg-core" containerID="cri-o://dbb316624926bb234d8295ddb2799f85b282e85bd10b8e727612c8fe23f64b7e" gracePeriod=30 Nov 28 11:45:04 crc kubenswrapper[4862]: I1128 11:45:04.457055 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="ceilometer-notification-agent" containerID="cri-o://f990be225595ff76feff0532ec85ced4559af58db51b56bec6908ff5195920c0" gracePeriod=30 Nov 28 11:45:05 crc kubenswrapper[4862]: I1128 11:45:05.267849 4862 generic.go:334] "Generic (PLEG): container finished" podID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerID="abd79235cd914e6150ae2d639121566a928bbb2d11a66e837754fa48b4877def" exitCode=0 Nov 28 11:45:05 crc kubenswrapper[4862]: I1128 11:45:05.267877 4862 generic.go:334] "Generic (PLEG): container finished" podID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerID="dbb316624926bb234d8295ddb2799f85b282e85bd10b8e727612c8fe23f64b7e" exitCode=2 Nov 28 11:45:05 crc kubenswrapper[4862]: I1128 11:45:05.267885 4862 generic.go:334] "Generic (PLEG): container finished" podID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerID="3bd7367ad3c8efcebb190a96352d52553a2cf7932124cdf7f8aa03f69869583a" exitCode=0 Nov 28 11:45:05 crc kubenswrapper[4862]: I1128 11:45:05.267906 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerDied","Data":"abd79235cd914e6150ae2d639121566a928bbb2d11a66e837754fa48b4877def"} Nov 28 11:45:05 crc kubenswrapper[4862]: I1128 11:45:05.267931 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerDied","Data":"dbb316624926bb234d8295ddb2799f85b282e85bd10b8e727612c8fe23f64b7e"} Nov 28 11:45:05 crc kubenswrapper[4862]: I1128 11:45:05.267940 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerDied","Data":"3bd7367ad3c8efcebb190a96352d52553a2cf7932124cdf7f8aa03f69869583a"} Nov 28 11:45:07 crc kubenswrapper[4862]: I1128 11:45:07.289154 4862 generic.go:334] "Generic (PLEG): container finished" podID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerID="f990be225595ff76feff0532ec85ced4559af58db51b56bec6908ff5195920c0" exitCode=0 Nov 28 11:45:07 crc kubenswrapper[4862]: I1128 11:45:07.289200 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerDied","Data":"f990be225595ff76feff0532ec85ced4559af58db51b56bec6908ff5195920c0"} Nov 28 11:45:07 crc kubenswrapper[4862]: I1128 11:45:07.551275 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:45:07 crc kubenswrapper[4862]: I1128 11:45:07.552381 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerName="glance-log" containerID="cri-o://ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95" gracePeriod=30 Nov 28 11:45:07 crc kubenswrapper[4862]: I1128 11:45:07.552964 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerName="glance-httpd" containerID="cri-o://b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba" gracePeriod=30 Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.169270 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.292000 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.292063 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.299542 4862 generic.go:334] "Generic (PLEG): container finished" podID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerID="ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95" exitCode=143 Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.299630 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"72c062a5-78c7-407a-8c0d-858a6fc301e2","Type":"ContainerDied","Data":"ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95"} Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.299803 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="635693e3-57cb-479e-b4b0-947f81f10e97" containerName="glance-log" containerID="cri-o://14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68" gracePeriod=30 Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.299887 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="635693e3-57cb-479e-b4b0-947f81f10e97" containerName="glance-httpd" containerID="cri-o://3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315" gracePeriod=30 Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.882618 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.896214 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99f601b7-1297-410b-b3e1-41bc2a1bb60b-secret-volume\") pod \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.896254 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99f601b7-1297-410b-b3e1-41bc2a1bb60b-config-volume\") pod \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.896353 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvjvx\" (UniqueName: \"kubernetes.io/projected/99f601b7-1297-410b-b3e1-41bc2a1bb60b-kube-api-access-lvjvx\") pod \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\" (UID: \"99f601b7-1297-410b-b3e1-41bc2a1bb60b\") " Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.904775 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99f601b7-1297-410b-b3e1-41bc2a1bb60b-kube-api-access-lvjvx" (OuterVolumeSpecName: "kube-api-access-lvjvx") pod "99f601b7-1297-410b-b3e1-41bc2a1bb60b" (UID: "99f601b7-1297-410b-b3e1-41bc2a1bb60b"). InnerVolumeSpecName "kube-api-access-lvjvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.907699 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99f601b7-1297-410b-b3e1-41bc2a1bb60b-config-volume" (OuterVolumeSpecName: "config-volume") pod "99f601b7-1297-410b-b3e1-41bc2a1bb60b" (UID: "99f601b7-1297-410b-b3e1-41bc2a1bb60b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:45:08 crc kubenswrapper[4862]: I1128 11:45:08.908728 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f601b7-1297-410b-b3e1-41bc2a1bb60b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "99f601b7-1297-410b-b3e1-41bc2a1bb60b" (UID: "99f601b7-1297-410b-b3e1-41bc2a1bb60b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:08.999369 4862 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99f601b7-1297-410b-b3e1-41bc2a1bb60b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:08.999408 4862 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99f601b7-1297-410b-b3e1-41bc2a1bb60b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:08.999422 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvjvx\" (UniqueName: \"kubernetes.io/projected/99f601b7-1297-410b-b3e1-41bc2a1bb60b-kube-api-access-lvjvx\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.144295 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.204293 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-config-data\") pod \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.204331 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-sg-core-conf-yaml\") pod \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.204351 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdtcz\" (UniqueName: \"kubernetes.io/projected/b26fee90-910d-4715-bdbd-4fe6d1695d4c-kube-api-access-jdtcz\") pod \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.204396 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-log-httpd\") pod \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.204489 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-run-httpd\") pod \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.204519 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-scripts\") pod \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.204538 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-combined-ca-bundle\") pod \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\" (UID: \"b26fee90-910d-4715-bdbd-4fe6d1695d4c\") " Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.207067 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b26fee90-910d-4715-bdbd-4fe6d1695d4c" (UID: "b26fee90-910d-4715-bdbd-4fe6d1695d4c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.207352 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b26fee90-910d-4715-bdbd-4fe6d1695d4c" (UID: "b26fee90-910d-4715-bdbd-4fe6d1695d4c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.209805 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b26fee90-910d-4715-bdbd-4fe6d1695d4c-kube-api-access-jdtcz" (OuterVolumeSpecName: "kube-api-access-jdtcz") pod "b26fee90-910d-4715-bdbd-4fe6d1695d4c" (UID: "b26fee90-910d-4715-bdbd-4fe6d1695d4c"). InnerVolumeSpecName "kube-api-access-jdtcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.219251 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-scripts" (OuterVolumeSpecName: "scripts") pod "b26fee90-910d-4715-bdbd-4fe6d1695d4c" (UID: "b26fee90-910d-4715-bdbd-4fe6d1695d4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.232172 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b26fee90-910d-4715-bdbd-4fe6d1695d4c" (UID: "b26fee90-910d-4715-bdbd-4fe6d1695d4c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.276938 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b26fee90-910d-4715-bdbd-4fe6d1695d4c" (UID: "b26fee90-910d-4715-bdbd-4fe6d1695d4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.306112 4862 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.306142 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdtcz\" (UniqueName: \"kubernetes.io/projected/b26fee90-910d-4715-bdbd-4fe6d1695d4c-kube-api-access-jdtcz\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.306153 4862 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.306163 4862 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26fee90-910d-4715-bdbd-4fe6d1695d4c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.306172 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.306182 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.311512 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2tm6f" event={"ID":"0cfe27ee-17c6-481a-81c0-6b48211db18c","Type":"ContainerStarted","Data":"72d37c2517d0efed5caf1404b498bb68525c5f4058d1148ee00d17548c09a99f"} Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.316495 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26fee90-910d-4715-bdbd-4fe6d1695d4c","Type":"ContainerDied","Data":"e52e0e1ae45980704418392b04dde3c69fcb570f0a0eef0e6d0bdad0da765d3b"} Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.316681 4862 scope.go:117] "RemoveContainer" containerID="abd79235cd914e6150ae2d639121566a928bbb2d11a66e837754fa48b4877def" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.316884 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.320158 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" event={"ID":"99f601b7-1297-410b-b3e1-41bc2a1bb60b","Type":"ContainerDied","Data":"97b6a34de1be80c3c3c0940b5678e22f51719982cc605a8f067eaefdae4431d1"} Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.320201 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97b6a34de1be80c3c3c0940b5678e22f51719982cc605a8f067eaefdae4431d1" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.320265 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405505-wf5ss" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.322201 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-config-data" (OuterVolumeSpecName: "config-data") pod "b26fee90-910d-4715-bdbd-4fe6d1695d4c" (UID: "b26fee90-910d-4715-bdbd-4fe6d1695d4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.330309 4862 generic.go:334] "Generic (PLEG): container finished" podID="635693e3-57cb-479e-b4b0-947f81f10e97" containerID="14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68" exitCode=143 Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.330363 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"635693e3-57cb-479e-b4b0-947f81f10e97","Type":"ContainerDied","Data":"14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68"} Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.340862 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-2tm6f" podStartSLOduration=1.619357551 podStartE2EDuration="11.340839062s" podCreationTimestamp="2025-11-28 11:44:58 +0000 UTC" firstStartedPulling="2025-11-28 11:44:58.997165141 +0000 UTC m=+1376.529679052" lastFinishedPulling="2025-11-28 11:45:08.718646642 +0000 UTC m=+1386.251160563" observedRunningTime="2025-11-28 11:45:09.330184745 +0000 UTC m=+1386.862698666" watchObservedRunningTime="2025-11-28 11:45:09.340839062 +0000 UTC m=+1386.873352983" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.342307 4862 scope.go:117] "RemoveContainer" containerID="dbb316624926bb234d8295ddb2799f85b282e85bd10b8e727612c8fe23f64b7e" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.363864 4862 scope.go:117] "RemoveContainer" containerID="f990be225595ff76feff0532ec85ced4559af58db51b56bec6908ff5195920c0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.391167 4862 scope.go:117] "RemoveContainer" containerID="3bd7367ad3c8efcebb190a96352d52553a2cf7932124cdf7f8aa03f69869583a" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.409619 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26fee90-910d-4715-bdbd-4fe6d1695d4c-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.696862 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.708510 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722198 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:09 crc kubenswrapper[4862]: E1128 11:45:09.722568 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="proxy-httpd" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722586 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="proxy-httpd" Nov 28 11:45:09 crc kubenswrapper[4862]: E1128 11:45:09.722600 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99f601b7-1297-410b-b3e1-41bc2a1bb60b" containerName="collect-profiles" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722606 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="99f601b7-1297-410b-b3e1-41bc2a1bb60b" containerName="collect-profiles" Nov 28 11:45:09 crc kubenswrapper[4862]: E1128 11:45:09.722624 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="sg-core" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722630 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="sg-core" Nov 28 11:45:09 crc kubenswrapper[4862]: E1128 11:45:09.722643 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="ceilometer-notification-agent" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722649 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="ceilometer-notification-agent" Nov 28 11:45:09 crc kubenswrapper[4862]: E1128 11:45:09.722668 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="ceilometer-central-agent" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722675 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="ceilometer-central-agent" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722839 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="ceilometer-central-agent" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722852 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="ceilometer-notification-agent" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722860 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="sg-core" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722875 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" containerName="proxy-httpd" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.722888 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="99f601b7-1297-410b-b3e1-41bc2a1bb60b" containerName="collect-profiles" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.724659 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.727007 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.727209 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.738040 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.816500 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-log-httpd\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.816788 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-run-httpd\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.817026 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.817074 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28djv\" (UniqueName: \"kubernetes.io/projected/cbc735a4-f48b-4d86-b8f5-15fd6b417750-kube-api-access-28djv\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.817440 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.817504 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-config-data\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.817534 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-scripts\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.920124 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-log-httpd\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.920204 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-run-httpd\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.920285 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.920319 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28djv\" (UniqueName: \"kubernetes.io/projected/cbc735a4-f48b-4d86-b8f5-15fd6b417750-kube-api-access-28djv\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.920490 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.920532 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-config-data\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.920570 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-scripts\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.922820 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-log-httpd\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.923329 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-run-httpd\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.937630 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.938783 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-config-data\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.938905 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-scripts\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.947680 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28djv\" (UniqueName: \"kubernetes.io/projected/cbc735a4-f48b-4d86-b8f5-15fd6b417750-kube-api-access-28djv\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:09 crc kubenswrapper[4862]: I1128 11:45:09.966913 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " pod="openstack/ceilometer-0" Nov 28 11:45:10 crc kubenswrapper[4862]: I1128 11:45:10.041397 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:45:10 crc kubenswrapper[4862]: I1128 11:45:10.491429 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:10 crc kubenswrapper[4862]: W1128 11:45:10.492735 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbc735a4_f48b_4d86_b8f5_15fd6b417750.slice/crio-01b51725cf47e644537995117e9216ae02387ad111e585cff1b1b1c2690e13db WatchSource:0}: Error finding container 01b51725cf47e644537995117e9216ae02387ad111e585cff1b1b1c2690e13db: Status 404 returned error can't find the container with id 01b51725cf47e644537995117e9216ae02387ad111e585cff1b1b1c2690e13db Nov 28 11:45:10 crc kubenswrapper[4862]: I1128 11:45:10.867186 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b26fee90-910d-4715-bdbd-4fe6d1695d4c" path="/var/lib/kubelet/pods/b26fee90-910d-4715-bdbd-4fe6d1695d4c/volumes" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.356524 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.362065 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerStarted","Data":"36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6"} Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.362120 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerStarted","Data":"01b51725cf47e644537995117e9216ae02387ad111e585cff1b1b1c2690e13db"} Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.367353 4862 generic.go:334] "Generic (PLEG): container finished" podID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerID="b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba" exitCode=0 Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.367385 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"72c062a5-78c7-407a-8c0d-858a6fc301e2","Type":"ContainerDied","Data":"b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba"} Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.367406 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"72c062a5-78c7-407a-8c0d-858a6fc301e2","Type":"ContainerDied","Data":"24f3b7c97ea9cdda492c58ab7f2634be3b89d39f189e7f18b24ce84fe9eeb385"} Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.367421 4862 scope.go:117] "RemoveContainer" containerID="b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.367529 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.418233 4862 scope.go:117] "RemoveContainer" containerID="ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.445685 4862 scope.go:117] "RemoveContainer" containerID="b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba" Nov 28 11:45:11 crc kubenswrapper[4862]: E1128 11:45:11.446258 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba\": container with ID starting with b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba not found: ID does not exist" containerID="b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.446312 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba"} err="failed to get container status \"b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba\": rpc error: code = NotFound desc = could not find container \"b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba\": container with ID starting with b0ac3370dafe64840722bb5e27731a5a12e462076a632a319fa83cacc5753fba not found: ID does not exist" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.446349 4862 scope.go:117] "RemoveContainer" containerID="ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95" Nov 28 11:45:11 crc kubenswrapper[4862]: E1128 11:45:11.446734 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95\": container with ID starting with ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95 not found: ID does not exist" containerID="ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.446827 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95"} err="failed to get container status \"ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95\": rpc error: code = NotFound desc = could not find container \"ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95\": container with ID starting with ee4260cf9c4d9fa22415d3f621f00fc51070a1b96450cb7be8014ad37293cf95 not found: ID does not exist" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.545808 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-scripts\") pod \"72c062a5-78c7-407a-8c0d-858a6fc301e2\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.545849 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-config-data\") pod \"72c062a5-78c7-407a-8c0d-858a6fc301e2\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.545868 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-public-tls-certs\") pod \"72c062a5-78c7-407a-8c0d-858a6fc301e2\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.545951 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-logs\") pod \"72c062a5-78c7-407a-8c0d-858a6fc301e2\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.545984 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-combined-ca-bundle\") pod \"72c062a5-78c7-407a-8c0d-858a6fc301e2\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.546681 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"72c062a5-78c7-407a-8c0d-858a6fc301e2\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.546742 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ckmp\" (UniqueName: \"kubernetes.io/projected/72c062a5-78c7-407a-8c0d-858a6fc301e2-kube-api-access-8ckmp\") pod \"72c062a5-78c7-407a-8c0d-858a6fc301e2\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.546769 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-httpd-run\") pod \"72c062a5-78c7-407a-8c0d-858a6fc301e2\" (UID: \"72c062a5-78c7-407a-8c0d-858a6fc301e2\") " Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.549708 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-logs" (OuterVolumeSpecName: "logs") pod "72c062a5-78c7-407a-8c0d-858a6fc301e2" (UID: "72c062a5-78c7-407a-8c0d-858a6fc301e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.551977 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "72c062a5-78c7-407a-8c0d-858a6fc301e2" (UID: "72c062a5-78c7-407a-8c0d-858a6fc301e2"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.552980 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-scripts" (OuterVolumeSpecName: "scripts") pod "72c062a5-78c7-407a-8c0d-858a6fc301e2" (UID: "72c062a5-78c7-407a-8c0d-858a6fc301e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.555152 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "72c062a5-78c7-407a-8c0d-858a6fc301e2" (UID: "72c062a5-78c7-407a-8c0d-858a6fc301e2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.563268 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72c062a5-78c7-407a-8c0d-858a6fc301e2-kube-api-access-8ckmp" (OuterVolumeSpecName: "kube-api-access-8ckmp") pod "72c062a5-78c7-407a-8c0d-858a6fc301e2" (UID: "72c062a5-78c7-407a-8c0d-858a6fc301e2"). InnerVolumeSpecName "kube-api-access-8ckmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.598011 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72c062a5-78c7-407a-8c0d-858a6fc301e2" (UID: "72c062a5-78c7-407a-8c0d-858a6fc301e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.619664 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-config-data" (OuterVolumeSpecName: "config-data") pod "72c062a5-78c7-407a-8c0d-858a6fc301e2" (UID: "72c062a5-78c7-407a-8c0d-858a6fc301e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.623370 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "72c062a5-78c7-407a-8c0d-858a6fc301e2" (UID: "72c062a5-78c7-407a-8c0d-858a6fc301e2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.649255 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.649547 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.649639 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.649720 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.649798 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c062a5-78c7-407a-8c0d-858a6fc301e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.649889 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.649968 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ckmp\" (UniqueName: \"kubernetes.io/projected/72c062a5-78c7-407a-8c0d-858a6fc301e2-kube-api-access-8ckmp\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.650043 4862 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72c062a5-78c7-407a-8c0d-858a6fc301e2-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.697737 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.746249 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.752004 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.756011 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.783198 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:45:11 crc kubenswrapper[4862]: E1128 11:45:11.786463 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerName="glance-httpd" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.786561 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerName="glance-httpd" Nov 28 11:45:11 crc kubenswrapper[4862]: E1128 11:45:11.786629 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerName="glance-log" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.786694 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerName="glance-log" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.786928 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerName="glance-log" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.786993 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c062a5-78c7-407a-8c0d-858a6fc301e2" containerName="glance-httpd" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.787942 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.790997 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.791254 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.816359 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.957761 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-logs\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.957815 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.957843 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-scripts\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.957959 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.958013 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rtl9\" (UniqueName: \"kubernetes.io/projected/e8ef59ee-556c-4735-8cdf-e140aa608f86-kube-api-access-6rtl9\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.958243 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-config-data\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.958329 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:11 crc kubenswrapper[4862]: I1128 11:45:11.958447 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.047947 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.060080 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.060195 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-logs\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.060234 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.060258 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-scripts\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.060341 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.060376 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rtl9\" (UniqueName: \"kubernetes.io/projected/e8ef59ee-556c-4735-8cdf-e140aa608f86-kube-api-access-6rtl9\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.060436 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-config-data\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.060479 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.060755 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.061346 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.061776 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-logs\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.066057 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-scripts\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.067192 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.068128 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.070685 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-config-data\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.100598 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rtl9\" (UniqueName: \"kubernetes.io/projected/e8ef59ee-556c-4735-8cdf-e140aa608f86-kube-api-access-6rtl9\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.129358 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.161240 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-internal-tls-certs\") pod \"635693e3-57cb-479e-b4b0-947f81f10e97\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.161314 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-combined-ca-bundle\") pod \"635693e3-57cb-479e-b4b0-947f81f10e97\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.161479 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-scripts\") pod \"635693e3-57cb-479e-b4b0-947f81f10e97\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.161525 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wc7f\" (UniqueName: \"kubernetes.io/projected/635693e3-57cb-479e-b4b0-947f81f10e97-kube-api-access-5wc7f\") pod \"635693e3-57cb-479e-b4b0-947f81f10e97\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.161590 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-httpd-run\") pod \"635693e3-57cb-479e-b4b0-947f81f10e97\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.161619 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-logs\") pod \"635693e3-57cb-479e-b4b0-947f81f10e97\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.162360 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-logs" (OuterVolumeSpecName: "logs") pod "635693e3-57cb-479e-b4b0-947f81f10e97" (UID: "635693e3-57cb-479e-b4b0-947f81f10e97"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.162720 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "635693e3-57cb-479e-b4b0-947f81f10e97" (UID: "635693e3-57cb-479e-b4b0-947f81f10e97"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.162809 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"635693e3-57cb-479e-b4b0-947f81f10e97\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.162852 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-config-data\") pod \"635693e3-57cb-479e-b4b0-947f81f10e97\" (UID: \"635693e3-57cb-479e-b4b0-947f81f10e97\") " Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.163399 4862 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.163416 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/635693e3-57cb-479e-b4b0-947f81f10e97-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.174780 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-scripts" (OuterVolumeSpecName: "scripts") pod "635693e3-57cb-479e-b4b0-947f81f10e97" (UID: "635693e3-57cb-479e-b4b0-947f81f10e97"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.175024 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/635693e3-57cb-479e-b4b0-947f81f10e97-kube-api-access-5wc7f" (OuterVolumeSpecName: "kube-api-access-5wc7f") pod "635693e3-57cb-479e-b4b0-947f81f10e97" (UID: "635693e3-57cb-479e-b4b0-947f81f10e97"). InnerVolumeSpecName "kube-api-access-5wc7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.180129 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "635693e3-57cb-479e-b4b0-947f81f10e97" (UID: "635693e3-57cb-479e-b4b0-947f81f10e97"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.247266 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "635693e3-57cb-479e-b4b0-947f81f10e97" (UID: "635693e3-57cb-479e-b4b0-947f81f10e97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.264605 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.264644 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.264657 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.264669 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wc7f\" (UniqueName: \"kubernetes.io/projected/635693e3-57cb-479e-b4b0-947f81f10e97-kube-api-access-5wc7f\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.279380 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-config-data" (OuterVolumeSpecName: "config-data") pod "635693e3-57cb-479e-b4b0-947f81f10e97" (UID: "635693e3-57cb-479e-b4b0-947f81f10e97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.289655 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.292060 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "635693e3-57cb-479e-b4b0-947f81f10e97" (UID: "635693e3-57cb-479e-b4b0-947f81f10e97"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.366166 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.366208 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/635693e3-57cb-479e-b4b0-947f81f10e97-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.366221 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.379011 4862 generic.go:334] "Generic (PLEG): container finished" podID="635693e3-57cb-479e-b4b0-947f81f10e97" containerID="3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315" exitCode=0 Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.379075 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"635693e3-57cb-479e-b4b0-947f81f10e97","Type":"ContainerDied","Data":"3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315"} Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.379124 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.379151 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"635693e3-57cb-479e-b4b0-947f81f10e97","Type":"ContainerDied","Data":"fc934e164af1fa4fee2ac658b0e1f1625f2caed49251d893dba2e1d0b7d26ae7"} Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.379185 4862 scope.go:117] "RemoveContainer" containerID="3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.385474 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerStarted","Data":"aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830"} Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.402934 4862 scope.go:117] "RemoveContainer" containerID="14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.427674 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.430478 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.440044 4862 scope.go:117] "RemoveContainer" containerID="3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315" Nov 28 11:45:12 crc kubenswrapper[4862]: E1128 11:45:12.444353 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315\": container with ID starting with 3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315 not found: ID does not exist" containerID="3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.444411 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315"} err="failed to get container status \"3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315\": rpc error: code = NotFound desc = could not find container \"3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315\": container with ID starting with 3f475f5c0bd6952c0f2cc59a06d9df8f3d8fd72322925cf63987ed8a674f2315 not found: ID does not exist" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.444446 4862 scope.go:117] "RemoveContainer" containerID="14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68" Nov 28 11:45:12 crc kubenswrapper[4862]: E1128 11:45:12.445146 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68\": container with ID starting with 14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68 not found: ID does not exist" containerID="14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.445175 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68"} err="failed to get container status \"14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68\": rpc error: code = NotFound desc = could not find container \"14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68\": container with ID starting with 14542ded85e6be3da70a2b3551eed97bb2b5a228721d2d7e014cfee0f3fd6d68 not found: ID does not exist" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.451621 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.461158 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:45:12 crc kubenswrapper[4862]: E1128 11:45:12.461593 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="635693e3-57cb-479e-b4b0-947f81f10e97" containerName="glance-httpd" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.461631 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="635693e3-57cb-479e-b4b0-947f81f10e97" containerName="glance-httpd" Nov 28 11:45:12 crc kubenswrapper[4862]: E1128 11:45:12.461671 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="635693e3-57cb-479e-b4b0-947f81f10e97" containerName="glance-log" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.461678 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="635693e3-57cb-479e-b4b0-947f81f10e97" containerName="glance-log" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.461839 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="635693e3-57cb-479e-b4b0-947f81f10e97" containerName="glance-httpd" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.461871 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="635693e3-57cb-479e-b4b0-947f81f10e97" containerName="glance-log" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.462807 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.465077 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.465338 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.523322 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.569494 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.569601 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.569646 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.569690 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26rfx\" (UniqueName: \"kubernetes.io/projected/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-kube-api-access-26rfx\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.569713 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.569793 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.569812 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.569841 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-logs\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.673548 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.673953 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.673973 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.674003 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-logs\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.674057 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.674140 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.674158 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.674198 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26rfx\" (UniqueName: \"kubernetes.io/projected/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-kube-api-access-26rfx\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.674767 4862 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.678717 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-logs\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.705465 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.707198 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.707796 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.717750 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.764694 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.860684 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.908520 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26rfx\" (UniqueName: \"kubernetes.io/projected/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-kube-api-access-26rfx\") pod \"glance-default-internal-api-0\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " pod="openstack/glance-default-internal-api-0" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.993258 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="635693e3-57cb-479e-b4b0-947f81f10e97" path="/var/lib/kubelet/pods/635693e3-57cb-479e-b4b0-947f81f10e97/volumes" Nov 28 11:45:12 crc kubenswrapper[4862]: I1128 11:45:12.994053 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72c062a5-78c7-407a-8c0d-858a6fc301e2" path="/var/lib/kubelet/pods/72c062a5-78c7-407a-8c0d-858a6fc301e2/volumes" Nov 28 11:45:13 crc kubenswrapper[4862]: I1128 11:45:13.134636 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:45:13 crc kubenswrapper[4862]: I1128 11:45:13.179384 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:13 crc kubenswrapper[4862]: I1128 11:45:13.420745 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerStarted","Data":"070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea"} Nov 28 11:45:13 crc kubenswrapper[4862]: I1128 11:45:13.423400 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e8ef59ee-556c-4735-8cdf-e140aa608f86","Type":"ContainerStarted","Data":"a99d9afba970fedc6cd31e5248110433094a405b5a46447f466d6177833cc195"} Nov 28 11:45:13 crc kubenswrapper[4862]: I1128 11:45:13.548456 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bbcf8e65-5e59-45e4-a5c3-f806fddd0c7a" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 11:45:13 crc kubenswrapper[4862]: I1128 11:45:13.612744 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:45:14 crc kubenswrapper[4862]: I1128 11:45:14.440613 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a","Type":"ContainerStarted","Data":"7e91caadb693c64c084738db969219970babbd935caf36c5d56c0b2e9a460f4a"} Nov 28 11:45:14 crc kubenswrapper[4862]: I1128 11:45:14.441324 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a","Type":"ContainerStarted","Data":"b332abc3cb712724bc373cf5882882de6d6b2dc3faa817a5caf4646ff373152b"} Nov 28 11:45:14 crc kubenswrapper[4862]: I1128 11:45:14.442920 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e8ef59ee-556c-4735-8cdf-e140aa608f86","Type":"ContainerStarted","Data":"236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb"} Nov 28 11:45:15 crc kubenswrapper[4862]: I1128 11:45:15.456361 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e8ef59ee-556c-4735-8cdf-e140aa608f86","Type":"ContainerStarted","Data":"ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8"} Nov 28 11:45:15 crc kubenswrapper[4862]: I1128 11:45:15.461482 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a","Type":"ContainerStarted","Data":"b622cbdd5687d59d52eeed34f730456b8b029c86f0df4d63c0373c42dc4992e7"} Nov 28 11:45:15 crc kubenswrapper[4862]: I1128 11:45:15.465520 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerStarted","Data":"93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb"} Nov 28 11:45:15 crc kubenswrapper[4862]: I1128 11:45:15.465725 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 11:45:15 crc kubenswrapper[4862]: I1128 11:45:15.497575 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.497542864 podStartE2EDuration="4.497542864s" podCreationTimestamp="2025-11-28 11:45:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:15.485476946 +0000 UTC m=+1393.017990867" watchObservedRunningTime="2025-11-28 11:45:15.497542864 +0000 UTC m=+1393.030056815" Nov 28 11:45:15 crc kubenswrapper[4862]: I1128 11:45:15.521634 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.5216108889999997 podStartE2EDuration="3.521610889s" podCreationTimestamp="2025-11-28 11:45:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:15.507374928 +0000 UTC m=+1393.039888849" watchObservedRunningTime="2025-11-28 11:45:15.521610889 +0000 UTC m=+1393.054124830" Nov 28 11:45:15 crc kubenswrapper[4862]: I1128 11:45:15.546499 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.458367349 podStartE2EDuration="6.546477716s" podCreationTimestamp="2025-11-28 11:45:09 +0000 UTC" firstStartedPulling="2025-11-28 11:45:10.494952837 +0000 UTC m=+1388.027466768" lastFinishedPulling="2025-11-28 11:45:14.583063214 +0000 UTC m=+1392.115577135" observedRunningTime="2025-11-28 11:45:15.53725869 +0000 UTC m=+1393.069772611" watchObservedRunningTime="2025-11-28 11:45:15.546477716 +0000 UTC m=+1393.078991637" Nov 28 11:45:21 crc kubenswrapper[4862]: I1128 11:45:21.542455 4862 generic.go:334] "Generic (PLEG): container finished" podID="0cfe27ee-17c6-481a-81c0-6b48211db18c" containerID="72d37c2517d0efed5caf1404b498bb68525c5f4058d1148ee00d17548c09a99f" exitCode=0 Nov 28 11:45:21 crc kubenswrapper[4862]: I1128 11:45:21.542542 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2tm6f" event={"ID":"0cfe27ee-17c6-481a-81c0-6b48211db18c","Type":"ContainerDied","Data":"72d37c2517d0efed5caf1404b498bb68525c5f4058d1148ee00d17548c09a99f"} Nov 28 11:45:22 crc kubenswrapper[4862]: I1128 11:45:22.431621 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 11:45:22 crc kubenswrapper[4862]: I1128 11:45:22.432073 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 11:45:22 crc kubenswrapper[4862]: I1128 11:45:22.483855 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 11:45:22 crc kubenswrapper[4862]: I1128 11:45:22.506946 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 11:45:22 crc kubenswrapper[4862]: I1128 11:45:22.569611 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 11:45:22 crc kubenswrapper[4862]: I1128 11:45:22.569702 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 11:45:22 crc kubenswrapper[4862]: I1128 11:45:22.990733 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.118968 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-scripts\") pod \"0cfe27ee-17c6-481a-81c0-6b48211db18c\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.119023 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-config-data\") pod \"0cfe27ee-17c6-481a-81c0-6b48211db18c\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.119049 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdxnq\" (UniqueName: \"kubernetes.io/projected/0cfe27ee-17c6-481a-81c0-6b48211db18c-kube-api-access-gdxnq\") pod \"0cfe27ee-17c6-481a-81c0-6b48211db18c\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.119151 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-combined-ca-bundle\") pod \"0cfe27ee-17c6-481a-81c0-6b48211db18c\" (UID: \"0cfe27ee-17c6-481a-81c0-6b48211db18c\") " Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.134850 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cfe27ee-17c6-481a-81c0-6b48211db18c-kube-api-access-gdxnq" (OuterVolumeSpecName: "kube-api-access-gdxnq") pod "0cfe27ee-17c6-481a-81c0-6b48211db18c" (UID: "0cfe27ee-17c6-481a-81c0-6b48211db18c"). InnerVolumeSpecName "kube-api-access-gdxnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.135321 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-scripts" (OuterVolumeSpecName: "scripts") pod "0cfe27ee-17c6-481a-81c0-6b48211db18c" (UID: "0cfe27ee-17c6-481a-81c0-6b48211db18c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.152691 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-config-data" (OuterVolumeSpecName: "config-data") pod "0cfe27ee-17c6-481a-81c0-6b48211db18c" (UID: "0cfe27ee-17c6-481a-81c0-6b48211db18c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.164140 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cfe27ee-17c6-481a-81c0-6b48211db18c" (UID: "0cfe27ee-17c6-481a-81c0-6b48211db18c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.180542 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.180598 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.222879 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.222932 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.222955 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdxnq\" (UniqueName: \"kubernetes.io/projected/0cfe27ee-17c6-481a-81c0-6b48211db18c-kube-api-access-gdxnq\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.222982 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cfe27ee-17c6-481a-81c0-6b48211db18c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.233516 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.241465 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.583259 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2tm6f" event={"ID":"0cfe27ee-17c6-481a-81c0-6b48211db18c","Type":"ContainerDied","Data":"f1a410ebb2478606e731dbad7b8280bddf0d97fee98531caad8715e0356f0410"} Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.583646 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1a410ebb2478606e731dbad7b8280bddf0d97fee98531caad8715e0356f0410" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.583675 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.583722 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.583395 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2tm6f" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.692291 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 11:45:23 crc kubenswrapper[4862]: E1128 11:45:23.692746 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfe27ee-17c6-481a-81c0-6b48211db18c" containerName="nova-cell0-conductor-db-sync" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.692768 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfe27ee-17c6-481a-81c0-6b48211db18c" containerName="nova-cell0-conductor-db-sync" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.692995 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfe27ee-17c6-481a-81c0-6b48211db18c" containerName="nova-cell0-conductor-db-sync" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.693742 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.698596 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.698829 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xdvln" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.739573 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.849497 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.849578 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.849607 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q5gq\" (UniqueName: \"kubernetes.io/projected/01f7b6ff-ee76-4961-9baf-930743da6143-kube-api-access-9q5gq\") pod \"nova-cell0-conductor-0\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.951749 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.951904 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.951951 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q5gq\" (UniqueName: \"kubernetes.io/projected/01f7b6ff-ee76-4961-9baf-930743da6143-kube-api-access-9q5gq\") pod \"nova-cell0-conductor-0\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.957698 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.966836 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:23 crc kubenswrapper[4862]: I1128 11:45:23.979391 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q5gq\" (UniqueName: \"kubernetes.io/projected/01f7b6ff-ee76-4961-9baf-930743da6143-kube-api-access-9q5gq\") pod \"nova-cell0-conductor-0\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:24 crc kubenswrapper[4862]: I1128 11:45:24.040716 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:24 crc kubenswrapper[4862]: I1128 11:45:24.579613 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 11:45:24 crc kubenswrapper[4862]: I1128 11:45:24.598839 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"01f7b6ff-ee76-4961-9baf-930743da6143","Type":"ContainerStarted","Data":"8d91061053a251e8e8dd0c4b382c70337c48209fecb8ce9c1d7b99feffd01a9e"} Nov 28 11:45:24 crc kubenswrapper[4862]: I1128 11:45:24.991550 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 11:45:24 crc kubenswrapper[4862]: I1128 11:45:24.991647 4862 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 11:45:24 crc kubenswrapper[4862]: I1128 11:45:24.999666 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 11:45:25 crc kubenswrapper[4862]: I1128 11:45:25.615677 4862 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 11:45:25 crc kubenswrapper[4862]: I1128 11:45:25.616038 4862 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 11:45:25 crc kubenswrapper[4862]: I1128 11:45:25.615863 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"01f7b6ff-ee76-4961-9baf-930743da6143","Type":"ContainerStarted","Data":"821e6798c6600a5f98070c95203ea984a4b500e28b5b89da8f188ebceafa93c7"} Nov 28 11:45:25 crc kubenswrapper[4862]: I1128 11:45:25.616548 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:25 crc kubenswrapper[4862]: I1128 11:45:25.641852 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.641834925 podStartE2EDuration="2.641834925s" podCreationTimestamp="2025-11-28 11:45:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:25.634704879 +0000 UTC m=+1403.167218890" watchObservedRunningTime="2025-11-28 11:45:25.641834925 +0000 UTC m=+1403.174348836" Nov 28 11:45:25 crc kubenswrapper[4862]: I1128 11:45:25.674959 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:25 crc kubenswrapper[4862]: I1128 11:45:25.808994 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.094086 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.630987 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-bnflp"] Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.632292 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.635821 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.636770 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.646329 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bnflp"] Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.767654 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.767810 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkx9v\" (UniqueName: \"kubernetes.io/projected/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-kube-api-access-wkx9v\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.767952 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-config-data\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.767989 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-scripts\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.830582 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.832060 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.843932 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.872933 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.873047 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkx9v\" (UniqueName: \"kubernetes.io/projected/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-kube-api-access-wkx9v\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.873114 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-config-data\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.873138 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-scripts\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.886048 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.887620 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-config-data\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.898731 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.902521 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-scripts\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.902782 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.904351 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.915719 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.939019 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.975590 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-config-data\") pod \"nova-scheduler-0\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.975695 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.975886 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzw8c\" (UniqueName: \"kubernetes.io/projected/153e6896-1533-4559-a5ba-b4bda00fac40-kube-api-access-zzw8c\") pod \"nova-scheduler-0\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.978015 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkx9v\" (UniqueName: \"kubernetes.io/projected/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-kube-api-access-wkx9v\") pod \"nova-cell0-cell-mapping-bnflp\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:34 crc kubenswrapper[4862]: I1128 11:45:34.981249 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.086168 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-config-data\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.086238 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5931c64-6372-4ef0-b916-5932a31a686e-logs\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.086306 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzw8c\" (UniqueName: \"kubernetes.io/projected/153e6896-1533-4559-a5ba-b4bda00fac40-kube-api-access-zzw8c\") pod \"nova-scheduler-0\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.086347 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-config-data\") pod \"nova-scheduler-0\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.086385 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.086419 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.086446 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpcjl\" (UniqueName: \"kubernetes.io/projected/e5931c64-6372-4ef0-b916-5932a31a686e-kube-api-access-vpcjl\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.102753 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-config-data\") pod \"nova-scheduler-0\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.104856 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.120472 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.122056 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.123624 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzw8c\" (UniqueName: \"kubernetes.io/projected/153e6896-1533-4559-a5ba-b4bda00fac40-kube-api-access-zzw8c\") pod \"nova-scheduler-0\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.127133 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.188045 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.188388 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpcjl\" (UniqueName: \"kubernetes.io/projected/e5931c64-6372-4ef0-b916-5932a31a686e-kube-api-access-vpcjl\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.188480 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-config-data\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.188571 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5931c64-6372-4ef0-b916-5932a31a686e-logs\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.190218 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5931c64-6372-4ef0-b916-5932a31a686e-logs\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.198828 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.199657 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75df6cf455-lxwbn"] Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.204006 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-config-data\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.207001 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.213959 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpcjl\" (UniqueName: \"kubernetes.io/projected/e5931c64-6372-4ef0-b916-5932a31a686e-kube-api-access-vpcjl\") pod \"nova-metadata-0\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.248625 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75df6cf455-lxwbn"] Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.255844 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.273208 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296127 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/678dbe9d-368e-48c2-9a21-42a18b2efeba-logs\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296214 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnlch\" (UniqueName: \"kubernetes.io/projected/c4edcc93-fcf3-46a4-afe4-0df67274a67f-kube-api-access-fnlch\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296229 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296263 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-config\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296320 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-config-data\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296349 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-swift-storage-0\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296391 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-sb\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296419 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-nb\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296447 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrm84\" (UniqueName: \"kubernetes.io/projected/678dbe9d-368e-48c2-9a21-42a18b2efeba-kube-api-access-zrm84\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296465 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.296146 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.299548 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.306342 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-svc\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.330662 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.357985 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415344 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/678dbe9d-368e-48c2-9a21-42a18b2efeba-logs\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415413 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnlch\" (UniqueName: \"kubernetes.io/projected/c4edcc93-fcf3-46a4-afe4-0df67274a67f-kube-api-access-fnlch\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415471 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-config\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415526 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-config-data\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415559 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-swift-storage-0\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415599 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-sb\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415629 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-nb\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415660 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrm84\" (UniqueName: \"kubernetes.io/projected/678dbe9d-368e-48c2-9a21-42a18b2efeba-kube-api-access-zrm84\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415682 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415735 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415781 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm7tw\" (UniqueName: \"kubernetes.io/projected/bad61764-bb4e-4bab-9712-e1b83597b642-kube-api-access-gm7tw\") pod \"nova-cell1-novncproxy-0\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.415805 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-svc\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.418221 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.419000 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/678dbe9d-368e-48c2-9a21-42a18b2efeba-logs\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.422521 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-nb\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.423057 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-config\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.429499 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-swift-storage-0\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.434188 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-svc\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.434777 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.435589 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-sb\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.457811 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-config-data\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.467334 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrm84\" (UniqueName: \"kubernetes.io/projected/678dbe9d-368e-48c2-9a21-42a18b2efeba-kube-api-access-zrm84\") pod \"nova-api-0\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.470913 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnlch\" (UniqueName: \"kubernetes.io/projected/c4edcc93-fcf3-46a4-afe4-0df67274a67f-kube-api-access-fnlch\") pod \"dnsmasq-dns-75df6cf455-lxwbn\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.525007 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.525103 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm7tw\" (UniqueName: \"kubernetes.io/projected/bad61764-bb4e-4bab-9712-e1b83597b642-kube-api-access-gm7tw\") pod \"nova-cell1-novncproxy-0\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.525134 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.536753 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.545653 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.560886 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm7tw\" (UniqueName: \"kubernetes.io/projected/bad61764-bb4e-4bab-9712-e1b83597b642-kube-api-access-gm7tw\") pod \"nova-cell1-novncproxy-0\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.586268 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.600678 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.638026 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:35 crc kubenswrapper[4862]: I1128 11:45:35.814820 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bnflp"] Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.049473 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:45:36 crc kubenswrapper[4862]: W1128 11:45:36.060798 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod678dbe9d_368e_48c2_9a21_42a18b2efeba.slice/crio-6113fffa459855f6bcf5bd0866d75172f1ec090299c4a8cf67ac8de8c737cf38 WatchSource:0}: Error finding container 6113fffa459855f6bcf5bd0866d75172f1ec090299c4a8cf67ac8de8c737cf38: Status 404 returned error can't find the container with id 6113fffa459855f6bcf5bd0866d75172f1ec090299c4a8cf67ac8de8c737cf38 Nov 28 11:45:36 crc kubenswrapper[4862]: W1128 11:45:36.072754 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod153e6896_1533_4559_a5ba_b4bda00fac40.slice/crio-de4bb2b9c942cda17adfeb34b6132e48aca782ecfe832881a1f91e3778bce288 WatchSource:0}: Error finding container de4bb2b9c942cda17adfeb34b6132e48aca782ecfe832881a1f91e3778bce288: Status 404 returned error can't find the container with id de4bb2b9c942cda17adfeb34b6132e48aca782ecfe832881a1f91e3778bce288 Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.079138 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.082755 4862 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.114144 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f8krc"] Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.115380 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.118675 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.118982 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.128732 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f8krc"] Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.143453 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.241543 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-config-data\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.242234 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-scripts\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.242311 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.242398 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bprnn\" (UniqueName: \"kubernetes.io/projected/0520c1cb-04b9-4790-93a1-25acd6925b94-kube-api-access-bprnn\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.301889 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75df6cf455-lxwbn"] Nov 28 11:45:36 crc kubenswrapper[4862]: W1128 11:45:36.307155 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4edcc93_fcf3_46a4_afe4_0df67274a67f.slice/crio-7a5eaf6915f3a5d400c748cfc7a14b5af21b248e596daf1ad5a3043c01903728 WatchSource:0}: Error finding container 7a5eaf6915f3a5d400c748cfc7a14b5af21b248e596daf1ad5a3043c01903728: Status 404 returned error can't find the container with id 7a5eaf6915f3a5d400c748cfc7a14b5af21b248e596daf1ad5a3043c01903728 Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.343792 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-config-data\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.343853 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-scripts\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.343930 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.343972 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bprnn\" (UniqueName: \"kubernetes.io/projected/0520c1cb-04b9-4790-93a1-25acd6925b94-kube-api-access-bprnn\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.349120 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.349541 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-scripts\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.349837 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-config-data\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.362724 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bprnn\" (UniqueName: \"kubernetes.io/projected/0520c1cb-04b9-4790-93a1-25acd6925b94-kube-api-access-bprnn\") pod \"nova-cell1-conductor-db-sync-f8krc\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.432920 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:45:36 crc kubenswrapper[4862]: W1128 11:45:36.438110 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbad61764_bb4e_4bab_9712_e1b83597b642.slice/crio-66c168311bc5f917103b0d6666120c76d65a9928a3f63a0b409df23adec739d8 WatchSource:0}: Error finding container 66c168311bc5f917103b0d6666120c76d65a9928a3f63a0b409df23adec739d8: Status 404 returned error can't find the container with id 66c168311bc5f917103b0d6666120c76d65a9928a3f63a0b409df23adec739d8 Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.493560 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.746663 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"678dbe9d-368e-48c2-9a21-42a18b2efeba","Type":"ContainerStarted","Data":"6113fffa459855f6bcf5bd0866d75172f1ec090299c4a8cf67ac8de8c737cf38"} Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.751043 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5931c64-6372-4ef0-b916-5932a31a686e","Type":"ContainerStarted","Data":"13e12733fa33049960d08c8f61b9f65b509c8199c5d57351750b8075c4a9994f"} Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.753930 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bnflp" event={"ID":"29592660-6aa1-41b7-a96d-ac6d3db4d7c1","Type":"ContainerStarted","Data":"a489662ec3258cf18713de72b1f2ca3ecbbd5ef2ee41d3665720ed4da7de42aa"} Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.753987 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bnflp" event={"ID":"29592660-6aa1-41b7-a96d-ac6d3db4d7c1","Type":"ContainerStarted","Data":"c0f092b0bd7f40ce706e00a3cefba7ff8caf82a1b29f39a3ac3d0a6fc5957c83"} Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.756822 4862 generic.go:334] "Generic (PLEG): container finished" podID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" containerID="a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076" exitCode=0 Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.756889 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" event={"ID":"c4edcc93-fcf3-46a4-afe4-0df67274a67f","Type":"ContainerDied","Data":"a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076"} Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.756913 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" event={"ID":"c4edcc93-fcf3-46a4-afe4-0df67274a67f","Type":"ContainerStarted","Data":"7a5eaf6915f3a5d400c748cfc7a14b5af21b248e596daf1ad5a3043c01903728"} Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.759017 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bad61764-bb4e-4bab-9712-e1b83597b642","Type":"ContainerStarted","Data":"66c168311bc5f917103b0d6666120c76d65a9928a3f63a0b409df23adec739d8"} Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.760217 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"153e6896-1533-4559-a5ba-b4bda00fac40","Type":"ContainerStarted","Data":"de4bb2b9c942cda17adfeb34b6132e48aca782ecfe832881a1f91e3778bce288"} Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.780503 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-bnflp" podStartSLOduration=2.780482179 podStartE2EDuration="2.780482179s" podCreationTimestamp="2025-11-28 11:45:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:36.778589405 +0000 UTC m=+1414.311103326" watchObservedRunningTime="2025-11-28 11:45:36.780482179 +0000 UTC m=+1414.312996100" Nov 28 11:45:36 crc kubenswrapper[4862]: I1128 11:45:36.961547 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f8krc"] Nov 28 11:45:36 crc kubenswrapper[4862]: W1128 11:45:36.969946 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0520c1cb_04b9_4790_93a1_25acd6925b94.slice/crio-573346c7b9a7e129e9b38a95b9cbddc0a55e8430a32343e983fa30af83380955 WatchSource:0}: Error finding container 573346c7b9a7e129e9b38a95b9cbddc0a55e8430a32343e983fa30af83380955: Status 404 returned error can't find the container with id 573346c7b9a7e129e9b38a95b9cbddc0a55e8430a32343e983fa30af83380955 Nov 28 11:45:37 crc kubenswrapper[4862]: I1128 11:45:37.773205 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f8krc" event={"ID":"0520c1cb-04b9-4790-93a1-25acd6925b94","Type":"ContainerStarted","Data":"369b84866b935e63fd85b445ef25d538f165e31d60f5a76ac7ab83b20dcfad0e"} Nov 28 11:45:37 crc kubenswrapper[4862]: I1128 11:45:37.773646 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f8krc" event={"ID":"0520c1cb-04b9-4790-93a1-25acd6925b94","Type":"ContainerStarted","Data":"573346c7b9a7e129e9b38a95b9cbddc0a55e8430a32343e983fa30af83380955"} Nov 28 11:45:37 crc kubenswrapper[4862]: I1128 11:45:37.779022 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" event={"ID":"c4edcc93-fcf3-46a4-afe4-0df67274a67f","Type":"ContainerStarted","Data":"3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c"} Nov 28 11:45:37 crc kubenswrapper[4862]: I1128 11:45:37.800182 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-f8krc" podStartSLOduration=1.8001621559999998 podStartE2EDuration="1.800162156s" podCreationTimestamp="2025-11-28 11:45:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:37.793130313 +0000 UTC m=+1415.325644234" watchObservedRunningTime="2025-11-28 11:45:37.800162156 +0000 UTC m=+1415.332676077" Nov 28 11:45:37 crc kubenswrapper[4862]: I1128 11:45:37.813210 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" podStartSLOduration=3.813192692 podStartE2EDuration="3.813192692s" podCreationTimestamp="2025-11-28 11:45:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:37.808642351 +0000 UTC m=+1415.341156272" watchObservedRunningTime="2025-11-28 11:45:37.813192692 +0000 UTC m=+1415.345706613" Nov 28 11:45:38 crc kubenswrapper[4862]: I1128 11:45:38.292381 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:45:38 crc kubenswrapper[4862]: I1128 11:45:38.292439 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:45:38 crc kubenswrapper[4862]: I1128 11:45:38.788163 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:39 crc kubenswrapper[4862]: I1128 11:45:39.255968 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:45:39 crc kubenswrapper[4862]: I1128 11:45:39.278193 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.046639 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.813831 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"678dbe9d-368e-48c2-9a21-42a18b2efeba","Type":"ContainerStarted","Data":"af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b"} Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.814185 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"678dbe9d-368e-48c2-9a21-42a18b2efeba","Type":"ContainerStarted","Data":"46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778"} Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.815867 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5931c64-6372-4ef0-b916-5932a31a686e","Type":"ContainerStarted","Data":"367fd41e1ba41032dbcd690e534b8d42bf4d31d68897baf42ad69746986c06f2"} Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.815906 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5931c64-6372-4ef0-b916-5932a31a686e","Type":"ContainerStarted","Data":"b303a2037bf85b8a831a2c5aef0ba62e1ac8700eb1a3b8afe9e0290837afcc2f"} Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.815941 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e5931c64-6372-4ef0-b916-5932a31a686e" containerName="nova-metadata-metadata" containerID="cri-o://367fd41e1ba41032dbcd690e534b8d42bf4d31d68897baf42ad69746986c06f2" gracePeriod=30 Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.815923 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e5931c64-6372-4ef0-b916-5932a31a686e" containerName="nova-metadata-log" containerID="cri-o://b303a2037bf85b8a831a2c5aef0ba62e1ac8700eb1a3b8afe9e0290837afcc2f" gracePeriod=30 Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.825854 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bad61764-bb4e-4bab-9712-e1b83597b642","Type":"ContainerStarted","Data":"dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca"} Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.825948 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="bad61764-bb4e-4bab-9712-e1b83597b642" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca" gracePeriod=30 Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.828410 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"153e6896-1533-4559-a5ba-b4bda00fac40","Type":"ContainerStarted","Data":"a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276"} Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.842281 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.373267311 podStartE2EDuration="6.842263567s" podCreationTimestamp="2025-11-28 11:45:34 +0000 UTC" firstStartedPulling="2025-11-28 11:45:36.082522285 +0000 UTC m=+1413.615036206" lastFinishedPulling="2025-11-28 11:45:39.551518541 +0000 UTC m=+1417.084032462" observedRunningTime="2025-11-28 11:45:40.83648328 +0000 UTC m=+1418.368997201" watchObservedRunningTime="2025-11-28 11:45:40.842263567 +0000 UTC m=+1418.374777488" Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.888394 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.765675381 podStartE2EDuration="5.888374767s" podCreationTimestamp="2025-11-28 11:45:35 +0000 UTC" firstStartedPulling="2025-11-28 11:45:36.444495307 +0000 UTC m=+1413.977009218" lastFinishedPulling="2025-11-28 11:45:39.567194673 +0000 UTC m=+1417.099708604" observedRunningTime="2025-11-28 11:45:40.86421194 +0000 UTC m=+1418.396725851" watchObservedRunningTime="2025-11-28 11:45:40.888374767 +0000 UTC m=+1418.420888688" Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.890505 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.435752063 podStartE2EDuration="6.890498118s" podCreationTimestamp="2025-11-28 11:45:34 +0000 UTC" firstStartedPulling="2025-11-28 11:45:36.095744496 +0000 UTC m=+1413.628258417" lastFinishedPulling="2025-11-28 11:45:39.550490541 +0000 UTC m=+1417.083004472" observedRunningTime="2025-11-28 11:45:40.88294778 +0000 UTC m=+1418.415461701" watchObservedRunningTime="2025-11-28 11:45:40.890498118 +0000 UTC m=+1418.423012039" Nov 28 11:45:40 crc kubenswrapper[4862]: I1128 11:45:40.906349 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.473818492 podStartE2EDuration="6.906335415s" podCreationTimestamp="2025-11-28 11:45:34 +0000 UTC" firstStartedPulling="2025-11-28 11:45:36.119183903 +0000 UTC m=+1413.651697834" lastFinishedPulling="2025-11-28 11:45:39.551700816 +0000 UTC m=+1417.084214757" observedRunningTime="2025-11-28 11:45:40.902741691 +0000 UTC m=+1418.435255612" watchObservedRunningTime="2025-11-28 11:45:40.906335415 +0000 UTC m=+1418.438849336" Nov 28 11:45:41 crc kubenswrapper[4862]: I1128 11:45:41.840209 4862 generic.go:334] "Generic (PLEG): container finished" podID="e5931c64-6372-4ef0-b916-5932a31a686e" containerID="367fd41e1ba41032dbcd690e534b8d42bf4d31d68897baf42ad69746986c06f2" exitCode=0 Nov 28 11:45:41 crc kubenswrapper[4862]: I1128 11:45:41.840574 4862 generic.go:334] "Generic (PLEG): container finished" podID="e5931c64-6372-4ef0-b916-5932a31a686e" containerID="b303a2037bf85b8a831a2c5aef0ba62e1ac8700eb1a3b8afe9e0290837afcc2f" exitCode=143 Nov 28 11:45:41 crc kubenswrapper[4862]: I1128 11:45:41.840292 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5931c64-6372-4ef0-b916-5932a31a686e","Type":"ContainerDied","Data":"367fd41e1ba41032dbcd690e534b8d42bf4d31d68897baf42ad69746986c06f2"} Nov 28 11:45:41 crc kubenswrapper[4862]: I1128 11:45:41.840622 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5931c64-6372-4ef0-b916-5932a31a686e","Type":"ContainerDied","Data":"b303a2037bf85b8a831a2c5aef0ba62e1ac8700eb1a3b8afe9e0290837afcc2f"} Nov 28 11:45:41 crc kubenswrapper[4862]: I1128 11:45:41.977790 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.067144 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-combined-ca-bundle\") pod \"e5931c64-6372-4ef0-b916-5932a31a686e\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.067206 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpcjl\" (UniqueName: \"kubernetes.io/projected/e5931c64-6372-4ef0-b916-5932a31a686e-kube-api-access-vpcjl\") pod \"e5931c64-6372-4ef0-b916-5932a31a686e\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.067263 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5931c64-6372-4ef0-b916-5932a31a686e-logs\") pod \"e5931c64-6372-4ef0-b916-5932a31a686e\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.067346 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-config-data\") pod \"e5931c64-6372-4ef0-b916-5932a31a686e\" (UID: \"e5931c64-6372-4ef0-b916-5932a31a686e\") " Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.075457 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5931c64-6372-4ef0-b916-5932a31a686e-logs" (OuterVolumeSpecName: "logs") pod "e5931c64-6372-4ef0-b916-5932a31a686e" (UID: "e5931c64-6372-4ef0-b916-5932a31a686e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.083776 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5931c64-6372-4ef0-b916-5932a31a686e-kube-api-access-vpcjl" (OuterVolumeSpecName: "kube-api-access-vpcjl") pod "e5931c64-6372-4ef0-b916-5932a31a686e" (UID: "e5931c64-6372-4ef0-b916-5932a31a686e"). InnerVolumeSpecName "kube-api-access-vpcjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.120525 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-config-data" (OuterVolumeSpecName: "config-data") pod "e5931c64-6372-4ef0-b916-5932a31a686e" (UID: "e5931c64-6372-4ef0-b916-5932a31a686e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.160329 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5931c64-6372-4ef0-b916-5932a31a686e" (UID: "e5931c64-6372-4ef0-b916-5932a31a686e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.171337 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.171371 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5931c64-6372-4ef0-b916-5932a31a686e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.171381 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpcjl\" (UniqueName: \"kubernetes.io/projected/e5931c64-6372-4ef0-b916-5932a31a686e-kube-api-access-vpcjl\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.171390 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5931c64-6372-4ef0-b916-5932a31a686e-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.856329 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.863911 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5931c64-6372-4ef0-b916-5932a31a686e","Type":"ContainerDied","Data":"13e12733fa33049960d08c8f61b9f65b509c8199c5d57351750b8075c4a9994f"} Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.863972 4862 scope.go:117] "RemoveContainer" containerID="367fd41e1ba41032dbcd690e534b8d42bf4d31d68897baf42ad69746986c06f2" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.914179 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.924917 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.940796 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:42 crc kubenswrapper[4862]: E1128 11:45:42.941203 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5931c64-6372-4ef0-b916-5932a31a686e" containerName="nova-metadata-log" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.941219 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5931c64-6372-4ef0-b916-5932a31a686e" containerName="nova-metadata-log" Nov 28 11:45:42 crc kubenswrapper[4862]: E1128 11:45:42.941261 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5931c64-6372-4ef0-b916-5932a31a686e" containerName="nova-metadata-metadata" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.941268 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5931c64-6372-4ef0-b916-5932a31a686e" containerName="nova-metadata-metadata" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.941465 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5931c64-6372-4ef0-b916-5932a31a686e" containerName="nova-metadata-log" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.941491 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5931c64-6372-4ef0-b916-5932a31a686e" containerName="nova-metadata-metadata" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.943421 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.955661 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.955868 4862 scope.go:117] "RemoveContainer" containerID="b303a2037bf85b8a831a2c5aef0ba62e1ac8700eb1a3b8afe9e0290837afcc2f" Nov 28 11:45:42 crc kubenswrapper[4862]: I1128 11:45:42.956002 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.005978 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.112146 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-config-data\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.112227 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa04a334-26a3-497b-9db3-9bae88388209-logs\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.112274 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lw4r\" (UniqueName: \"kubernetes.io/projected/aa04a334-26a3-497b-9db3-9bae88388209-kube-api-access-2lw4r\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.112497 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.112613 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.213832 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa04a334-26a3-497b-9db3-9bae88388209-logs\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.214151 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lw4r\" (UniqueName: \"kubernetes.io/projected/aa04a334-26a3-497b-9db3-9bae88388209-kube-api-access-2lw4r\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.214219 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.214252 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.214297 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-config-data\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.214642 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa04a334-26a3-497b-9db3-9bae88388209-logs\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.220000 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-config-data\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.221819 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.222581 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.234761 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lw4r\" (UniqueName: \"kubernetes.io/projected/aa04a334-26a3-497b-9db3-9bae88388209-kube-api-access-2lw4r\") pod \"nova-metadata-0\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.282635 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:43 crc kubenswrapper[4862]: W1128 11:45:43.596308 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa04a334_26a3_497b_9db3_9bae88388209.slice/crio-088e7ceab406db32323def7202bbc8ee8a1b2464708a07250cdc02e3d21e4453 WatchSource:0}: Error finding container 088e7ceab406db32323def7202bbc8ee8a1b2464708a07250cdc02e3d21e4453: Status 404 returned error can't find the container with id 088e7ceab406db32323def7202bbc8ee8a1b2464708a07250cdc02e3d21e4453 Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.603045 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.866701 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa04a334-26a3-497b-9db3-9bae88388209","Type":"ContainerStarted","Data":"13ea66c81ff754ef5ada0de2b9353279bd3f56a0dc6a739590c85183e537da8e"} Nov 28 11:45:43 crc kubenswrapper[4862]: I1128 11:45:43.867077 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa04a334-26a3-497b-9db3-9bae88388209","Type":"ContainerStarted","Data":"088e7ceab406db32323def7202bbc8ee8a1b2464708a07250cdc02e3d21e4453"} Nov 28 11:45:44 crc kubenswrapper[4862]: I1128 11:45:44.831824 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:45:44 crc kubenswrapper[4862]: I1128 11:45:44.832249 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="264bc03e-9ff0-44ce-b7d3-05111d5b6483" containerName="kube-state-metrics" containerID="cri-o://c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11" gracePeriod=30 Nov 28 11:45:44 crc kubenswrapper[4862]: I1128 11:45:44.849622 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5931c64-6372-4ef0-b916-5932a31a686e" path="/var/lib/kubelet/pods/e5931c64-6372-4ef0-b916-5932a31a686e/volumes" Nov 28 11:45:44 crc kubenswrapper[4862]: I1128 11:45:44.894254 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa04a334-26a3-497b-9db3-9bae88388209","Type":"ContainerStarted","Data":"61cfcc4fce39fe3c82b8a0422d64efcda70ec22171fe40b6fdd21c27dbcf7bd7"} Nov 28 11:45:44 crc kubenswrapper[4862]: I1128 11:45:44.921263 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.92124856 podStartE2EDuration="2.92124856s" podCreationTimestamp="2025-11-28 11:45:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:44.912412345 +0000 UTC m=+1422.444926266" watchObservedRunningTime="2025-11-28 11:45:44.92124856 +0000 UTC m=+1422.453762471" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.332665 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.332711 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.385750 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.567810 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.587235 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.587285 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.604261 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.638432 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.674625 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c55f6679-bknmm"] Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.674844 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c55f6679-bknmm" podUID="d61f3392-fcc7-42dd-8028-172d4f5e04b2" containerName="dnsmasq-dns" containerID="cri-o://bde25c7a11d728f465b47e43d012c44bfe12d09c2b631dad13c8f797360f6927" gracePeriod=10 Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.685030 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ns99\" (UniqueName: \"kubernetes.io/projected/264bc03e-9ff0-44ce-b7d3-05111d5b6483-kube-api-access-4ns99\") pod \"264bc03e-9ff0-44ce-b7d3-05111d5b6483\" (UID: \"264bc03e-9ff0-44ce-b7d3-05111d5b6483\") " Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.699517 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/264bc03e-9ff0-44ce-b7d3-05111d5b6483-kube-api-access-4ns99" (OuterVolumeSpecName: "kube-api-access-4ns99") pod "264bc03e-9ff0-44ce-b7d3-05111d5b6483" (UID: "264bc03e-9ff0-44ce-b7d3-05111d5b6483"). InnerVolumeSpecName "kube-api-access-4ns99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.787879 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ns99\" (UniqueName: \"kubernetes.io/projected/264bc03e-9ff0-44ce-b7d3-05111d5b6483-kube-api-access-4ns99\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.931311 4862 generic.go:334] "Generic (PLEG): container finished" podID="264bc03e-9ff0-44ce-b7d3-05111d5b6483" containerID="c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11" exitCode=2 Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.931437 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"264bc03e-9ff0-44ce-b7d3-05111d5b6483","Type":"ContainerDied","Data":"c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11"} Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.931465 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"264bc03e-9ff0-44ce-b7d3-05111d5b6483","Type":"ContainerDied","Data":"b3246f0ffa7717c844ee023f89483fdbb0fe8058f206dcf92d6188b0fc6d998c"} Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.931486 4862 scope.go:117] "RemoveContainer" containerID="c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.931666 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.942187 4862 generic.go:334] "Generic (PLEG): container finished" podID="29592660-6aa1-41b7-a96d-ac6d3db4d7c1" containerID="a489662ec3258cf18713de72b1f2ca3ecbbd5ef2ee41d3665720ed4da7de42aa" exitCode=0 Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.942256 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bnflp" event={"ID":"29592660-6aa1-41b7-a96d-ac6d3db4d7c1","Type":"ContainerDied","Data":"a489662ec3258cf18713de72b1f2ca3ecbbd5ef2ee41d3665720ed4da7de42aa"} Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.954340 4862 generic.go:334] "Generic (PLEG): container finished" podID="0520c1cb-04b9-4790-93a1-25acd6925b94" containerID="369b84866b935e63fd85b445ef25d538f165e31d60f5a76ac7ab83b20dcfad0e" exitCode=0 Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.954403 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f8krc" event={"ID":"0520c1cb-04b9-4790-93a1-25acd6925b94","Type":"ContainerDied","Data":"369b84866b935e63fd85b445ef25d538f165e31d60f5a76ac7ab83b20dcfad0e"} Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.966924 4862 scope.go:117] "RemoveContainer" containerID="c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.968716 4862 generic.go:334] "Generic (PLEG): container finished" podID="d61f3392-fcc7-42dd-8028-172d4f5e04b2" containerID="bde25c7a11d728f465b47e43d012c44bfe12d09c2b631dad13c8f797360f6927" exitCode=0 Nov 28 11:45:45 crc kubenswrapper[4862]: E1128 11:45:45.969892 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11\": container with ID starting with c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11 not found: ID does not exist" containerID="c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11" Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.969923 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c55f6679-bknmm" event={"ID":"d61f3392-fcc7-42dd-8028-172d4f5e04b2","Type":"ContainerDied","Data":"bde25c7a11d728f465b47e43d012c44bfe12d09c2b631dad13c8f797360f6927"} Nov 28 11:45:45 crc kubenswrapper[4862]: I1128 11:45:45.969943 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11"} err="failed to get container status \"c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11\": rpc error: code = NotFound desc = could not find container \"c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11\": container with ID starting with c38c78909b9b74539cd47ab3188f32949f047cb22afb40c5ff59770c5928aa11 not found: ID does not exist" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.003323 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.016670 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.019079 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.029241 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:45:46 crc kubenswrapper[4862]: E1128 11:45:46.029810 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264bc03e-9ff0-44ce-b7d3-05111d5b6483" containerName="kube-state-metrics" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.029826 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="264bc03e-9ff0-44ce-b7d3-05111d5b6483" containerName="kube-state-metrics" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.030109 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="264bc03e-9ff0-44ce-b7d3-05111d5b6483" containerName="kube-state-metrics" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.033604 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.036074 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.036222 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.039440 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.198387 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.198611 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.198740 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.198783 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6hmh\" (UniqueName: \"kubernetes.io/projected/164085a5-ab64-4c9b-b144-238e35ab3219-kube-api-access-q6hmh\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.200235 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.300691 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-svc\") pod \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.300759 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-swift-storage-0\") pod \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.300949 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-config\") pod \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.301002 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-nb\") pod \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.301030 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-sb\") pod \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.301062 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmts9\" (UniqueName: \"kubernetes.io/projected/d61f3392-fcc7-42dd-8028-172d4f5e04b2-kube-api-access-vmts9\") pod \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\" (UID: \"d61f3392-fcc7-42dd-8028-172d4f5e04b2\") " Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.301443 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.301545 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.301611 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.301641 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6hmh\" (UniqueName: \"kubernetes.io/projected/164085a5-ab64-4c9b-b144-238e35ab3219-kube-api-access-q6hmh\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.310712 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.310919 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.317870 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.318157 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d61f3392-fcc7-42dd-8028-172d4f5e04b2-kube-api-access-vmts9" (OuterVolumeSpecName: "kube-api-access-vmts9") pod "d61f3392-fcc7-42dd-8028-172d4f5e04b2" (UID: "d61f3392-fcc7-42dd-8028-172d4f5e04b2"). InnerVolumeSpecName "kube-api-access-vmts9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.325037 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6hmh\" (UniqueName: \"kubernetes.io/projected/164085a5-ab64-4c9b-b144-238e35ab3219-kube-api-access-q6hmh\") pod \"kube-state-metrics-0\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.360824 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.372733 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d61f3392-fcc7-42dd-8028-172d4f5e04b2" (UID: "d61f3392-fcc7-42dd-8028-172d4f5e04b2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.384758 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d61f3392-fcc7-42dd-8028-172d4f5e04b2" (UID: "d61f3392-fcc7-42dd-8028-172d4f5e04b2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.384773 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d61f3392-fcc7-42dd-8028-172d4f5e04b2" (UID: "d61f3392-fcc7-42dd-8028-172d4f5e04b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.389172 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d61f3392-fcc7-42dd-8028-172d4f5e04b2" (UID: "d61f3392-fcc7-42dd-8028-172d4f5e04b2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.390744 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-config" (OuterVolumeSpecName: "config") pod "d61f3392-fcc7-42dd-8028-172d4f5e04b2" (UID: "d61f3392-fcc7-42dd-8028-172d4f5e04b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.403604 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.403625 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.403633 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.403643 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmts9\" (UniqueName: \"kubernetes.io/projected/d61f3392-fcc7-42dd-8028-172d4f5e04b2-kube-api-access-vmts9\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.403651 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.403661 4862 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d61f3392-fcc7-42dd-8028-172d4f5e04b2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.671505 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.671582 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.852764 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="264bc03e-9ff0-44ce-b7d3-05111d5b6483" path="/var/lib/kubelet/pods/264bc03e-9ff0-44ce-b7d3-05111d5b6483/volumes" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.872539 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.978613 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c55f6679-bknmm" event={"ID":"d61f3392-fcc7-42dd-8028-172d4f5e04b2","Type":"ContainerDied","Data":"bd5f19d9ff387922676603cab18fa3ece93be4cab7adce347b391e3bdcf2d8f1"} Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.978677 4862 scope.go:117] "RemoveContainer" containerID="bde25c7a11d728f465b47e43d012c44bfe12d09c2b631dad13c8f797360f6927" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.978808 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c55f6679-bknmm" Nov 28 11:45:46 crc kubenswrapper[4862]: I1128 11:45:46.981328 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"164085a5-ab64-4c9b-b144-238e35ab3219","Type":"ContainerStarted","Data":"d4d25375805a2c1a55dbdedaa7447bc1b0e831a818fc3da9ce359c0fb27a9eab"} Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.039252 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c55f6679-bknmm"] Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.046435 4862 scope.go:117] "RemoveContainer" containerID="6a3cd0f4a9e484c02e4255709623e45cdfd370830a5817ea1e8c645d567544be" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.056844 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c55f6679-bknmm"] Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.440506 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.441043 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="ceilometer-central-agent" containerID="cri-o://36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6" gracePeriod=30 Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.441740 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="proxy-httpd" containerID="cri-o://93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb" gracePeriod=30 Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.441788 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="sg-core" containerID="cri-o://070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea" gracePeriod=30 Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.441822 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="ceilometer-notification-agent" containerID="cri-o://aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830" gracePeriod=30 Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.473898 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.529618 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-config-data\") pod \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.529828 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-scripts\") pod \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.529949 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-combined-ca-bundle\") pod \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.530141 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkx9v\" (UniqueName: \"kubernetes.io/projected/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-kube-api-access-wkx9v\") pod \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\" (UID: \"29592660-6aa1-41b7-a96d-ac6d3db4d7c1\") " Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.535385 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-scripts" (OuterVolumeSpecName: "scripts") pod "29592660-6aa1-41b7-a96d-ac6d3db4d7c1" (UID: "29592660-6aa1-41b7-a96d-ac6d3db4d7c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.573261 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-kube-api-access-wkx9v" (OuterVolumeSpecName: "kube-api-access-wkx9v") pod "29592660-6aa1-41b7-a96d-ac6d3db4d7c1" (UID: "29592660-6aa1-41b7-a96d-ac6d3db4d7c1"). InnerVolumeSpecName "kube-api-access-wkx9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.587912 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29592660-6aa1-41b7-a96d-ac6d3db4d7c1" (UID: "29592660-6aa1-41b7-a96d-ac6d3db4d7c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.592736 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-config-data" (OuterVolumeSpecName: "config-data") pod "29592660-6aa1-41b7-a96d-ac6d3db4d7c1" (UID: "29592660-6aa1-41b7-a96d-ac6d3db4d7c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.633890 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.633998 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.634458 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkx9v\" (UniqueName: \"kubernetes.io/projected/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-kube-api-access-wkx9v\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.634525 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29592660-6aa1-41b7-a96d-ac6d3db4d7c1-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.658366 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.735252 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-scripts\") pod \"0520c1cb-04b9-4790-93a1-25acd6925b94\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.735306 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-combined-ca-bundle\") pod \"0520c1cb-04b9-4790-93a1-25acd6925b94\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.735404 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-config-data\") pod \"0520c1cb-04b9-4790-93a1-25acd6925b94\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.735460 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bprnn\" (UniqueName: \"kubernetes.io/projected/0520c1cb-04b9-4790-93a1-25acd6925b94-kube-api-access-bprnn\") pod \"0520c1cb-04b9-4790-93a1-25acd6925b94\" (UID: \"0520c1cb-04b9-4790-93a1-25acd6925b94\") " Nov 28 11:45:47 crc kubenswrapper[4862]: E1128 11:45:47.738888 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbc735a4_f48b_4d86_b8f5_15fd6b417750.slice/crio-070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbc735a4_f48b_4d86_b8f5_15fd6b417750.slice/crio-93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb.scope\": RecentStats: unable to find data in memory cache]" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.760192 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-scripts" (OuterVolumeSpecName: "scripts") pod "0520c1cb-04b9-4790-93a1-25acd6925b94" (UID: "0520c1cb-04b9-4790-93a1-25acd6925b94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.760457 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0520c1cb-04b9-4790-93a1-25acd6925b94-kube-api-access-bprnn" (OuterVolumeSpecName: "kube-api-access-bprnn") pod "0520c1cb-04b9-4790-93a1-25acd6925b94" (UID: "0520c1cb-04b9-4790-93a1-25acd6925b94"). InnerVolumeSpecName "kube-api-access-bprnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.769944 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0520c1cb-04b9-4790-93a1-25acd6925b94" (UID: "0520c1cb-04b9-4790-93a1-25acd6925b94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.785361 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-config-data" (OuterVolumeSpecName: "config-data") pod "0520c1cb-04b9-4790-93a1-25acd6925b94" (UID: "0520c1cb-04b9-4790-93a1-25acd6925b94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.837890 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.837921 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.837933 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0520c1cb-04b9-4790-93a1-25acd6925b94-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.837949 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bprnn\" (UniqueName: \"kubernetes.io/projected/0520c1cb-04b9-4790-93a1-25acd6925b94-kube-api-access-bprnn\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.999705 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f8krc" event={"ID":"0520c1cb-04b9-4790-93a1-25acd6925b94","Type":"ContainerDied","Data":"573346c7b9a7e129e9b38a95b9cbddc0a55e8430a32343e983fa30af83380955"} Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.999757 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="573346c7b9a7e129e9b38a95b9cbddc0a55e8430a32343e983fa30af83380955" Nov 28 11:45:47 crc kubenswrapper[4862]: I1128 11:45:47.999824 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f8krc" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.010538 4862 generic.go:334] "Generic (PLEG): container finished" podID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerID="93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb" exitCode=0 Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.010590 4862 generic.go:334] "Generic (PLEG): container finished" podID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerID="070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea" exitCode=2 Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.010601 4862 generic.go:334] "Generic (PLEG): container finished" podID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerID="36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6" exitCode=0 Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.010660 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerDied","Data":"93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb"} Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.010694 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerDied","Data":"070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea"} Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.010708 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerDied","Data":"36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6"} Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.021722 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"164085a5-ab64-4c9b-b144-238e35ab3219","Type":"ContainerStarted","Data":"16fc0902091456b3028a91ab947a0a28c767f3d15650008ec782b9684a580f92"} Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.021894 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.028788 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bnflp" event={"ID":"29592660-6aa1-41b7-a96d-ac6d3db4d7c1","Type":"ContainerDied","Data":"c0f092b0bd7f40ce706e00a3cefba7ff8caf82a1b29f39a3ac3d0a6fc5957c83"} Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.028826 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0f092b0bd7f40ce706e00a3cefba7ff8caf82a1b29f39a3ac3d0a6fc5957c83" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.028885 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bnflp" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.053133 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.648357443 podStartE2EDuration="3.05311683s" podCreationTimestamp="2025-11-28 11:45:45 +0000 UTC" firstStartedPulling="2025-11-28 11:45:46.884556909 +0000 UTC m=+1424.417070840" lastFinishedPulling="2025-11-28 11:45:47.289316286 +0000 UTC m=+1424.821830227" observedRunningTime="2025-11-28 11:45:48.042190615 +0000 UTC m=+1425.574704526" watchObservedRunningTime="2025-11-28 11:45:48.05311683 +0000 UTC m=+1425.585630741" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.087374 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 11:45:48 crc kubenswrapper[4862]: E1128 11:45:48.087848 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29592660-6aa1-41b7-a96d-ac6d3db4d7c1" containerName="nova-manage" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.087864 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="29592660-6aa1-41b7-a96d-ac6d3db4d7c1" containerName="nova-manage" Nov 28 11:45:48 crc kubenswrapper[4862]: E1128 11:45:48.087889 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0520c1cb-04b9-4790-93a1-25acd6925b94" containerName="nova-cell1-conductor-db-sync" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.087898 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="0520c1cb-04b9-4790-93a1-25acd6925b94" containerName="nova-cell1-conductor-db-sync" Nov 28 11:45:48 crc kubenswrapper[4862]: E1128 11:45:48.087917 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61f3392-fcc7-42dd-8028-172d4f5e04b2" containerName="init" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.087925 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61f3392-fcc7-42dd-8028-172d4f5e04b2" containerName="init" Nov 28 11:45:48 crc kubenswrapper[4862]: E1128 11:45:48.087963 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61f3392-fcc7-42dd-8028-172d4f5e04b2" containerName="dnsmasq-dns" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.087970 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61f3392-fcc7-42dd-8028-172d4f5e04b2" containerName="dnsmasq-dns" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.088219 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="d61f3392-fcc7-42dd-8028-172d4f5e04b2" containerName="dnsmasq-dns" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.088230 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="29592660-6aa1-41b7-a96d-ac6d3db4d7c1" containerName="nova-manage" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.088250 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="0520c1cb-04b9-4790-93a1-25acd6925b94" containerName="nova-cell1-conductor-db-sync" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.088862 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.092445 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.103336 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.182355 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.183503 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-api" containerID="cri-o://af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b" gracePeriod=30 Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.183664 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-log" containerID="cri-o://46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778" gracePeriod=30 Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.206690 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.206976 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="153e6896-1533-4559-a5ba-b4bda00fac40" containerName="nova-scheduler-scheduler" containerID="cri-o://a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276" gracePeriod=30 Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.217012 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.217310 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="aa04a334-26a3-497b-9db3-9bae88388209" containerName="nova-metadata-metadata" containerID="cri-o://61cfcc4fce39fe3c82b8a0422d64efcda70ec22171fe40b6fdd21c27dbcf7bd7" gracePeriod=30 Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.217712 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="aa04a334-26a3-497b-9db3-9bae88388209" containerName="nova-metadata-log" containerID="cri-o://13ea66c81ff754ef5ada0de2b9353279bd3f56a0dc6a739590c85183e537da8e" gracePeriod=30 Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.259674 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.259765 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.259787 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xqpn\" (UniqueName: \"kubernetes.io/projected/ae02e889-f98f-4678-b15f-d91f711083f9-kube-api-access-8xqpn\") pod \"nova-cell1-conductor-0\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.284418 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.284496 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.361620 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.361680 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xqpn\" (UniqueName: \"kubernetes.io/projected/ae02e889-f98f-4678-b15f-d91f711083f9-kube-api-access-8xqpn\") pod \"nova-cell1-conductor-0\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.361838 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.367074 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.370034 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.386734 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xqpn\" (UniqueName: \"kubernetes.io/projected/ae02e889-f98f-4678-b15f-d91f711083f9-kube-api-access-8xqpn\") pod \"nova-cell1-conductor-0\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.408500 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.852568 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d61f3392-fcc7-42dd-8028-172d4f5e04b2" path="/var/lib/kubelet/pods/d61f3392-fcc7-42dd-8028-172d4f5e04b2/volumes" Nov 28 11:45:48 crc kubenswrapper[4862]: I1128 11:45:48.991899 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.041654 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ae02e889-f98f-4678-b15f-d91f711083f9","Type":"ContainerStarted","Data":"e1e0952a769d2d2ea44aeded428a163cea3def05adf8bb7b2dcb71f595cd6959"} Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.049357 4862 generic.go:334] "Generic (PLEG): container finished" podID="aa04a334-26a3-497b-9db3-9bae88388209" containerID="61cfcc4fce39fe3c82b8a0422d64efcda70ec22171fe40b6fdd21c27dbcf7bd7" exitCode=0 Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.049399 4862 generic.go:334] "Generic (PLEG): container finished" podID="aa04a334-26a3-497b-9db3-9bae88388209" containerID="13ea66c81ff754ef5ada0de2b9353279bd3f56a0dc6a739590c85183e537da8e" exitCode=143 Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.049443 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa04a334-26a3-497b-9db3-9bae88388209","Type":"ContainerDied","Data":"61cfcc4fce39fe3c82b8a0422d64efcda70ec22171fe40b6fdd21c27dbcf7bd7"} Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.049664 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa04a334-26a3-497b-9db3-9bae88388209","Type":"ContainerDied","Data":"13ea66c81ff754ef5ada0de2b9353279bd3f56a0dc6a739590c85183e537da8e"} Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.049687 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa04a334-26a3-497b-9db3-9bae88388209","Type":"ContainerDied","Data":"088e7ceab406db32323def7202bbc8ee8a1b2464708a07250cdc02e3d21e4453"} Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.049697 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="088e7ceab406db32323def7202bbc8ee8a1b2464708a07250cdc02e3d21e4453" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.051566 4862 generic.go:334] "Generic (PLEG): container finished" podID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerID="46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778" exitCode=143 Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.052423 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"678dbe9d-368e-48c2-9a21-42a18b2efeba","Type":"ContainerDied","Data":"46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778"} Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.139419 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.293679 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-config-data\") pod \"aa04a334-26a3-497b-9db3-9bae88388209\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.293734 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa04a334-26a3-497b-9db3-9bae88388209-logs\") pod \"aa04a334-26a3-497b-9db3-9bae88388209\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.293756 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-nova-metadata-tls-certs\") pod \"aa04a334-26a3-497b-9db3-9bae88388209\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.293813 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lw4r\" (UniqueName: \"kubernetes.io/projected/aa04a334-26a3-497b-9db3-9bae88388209-kube-api-access-2lw4r\") pod \"aa04a334-26a3-497b-9db3-9bae88388209\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.293914 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-combined-ca-bundle\") pod \"aa04a334-26a3-497b-9db3-9bae88388209\" (UID: \"aa04a334-26a3-497b-9db3-9bae88388209\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.295510 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa04a334-26a3-497b-9db3-9bae88388209-logs" (OuterVolumeSpecName: "logs") pod "aa04a334-26a3-497b-9db3-9bae88388209" (UID: "aa04a334-26a3-497b-9db3-9bae88388209"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.300002 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa04a334-26a3-497b-9db3-9bae88388209-kube-api-access-2lw4r" (OuterVolumeSpecName: "kube-api-access-2lw4r") pod "aa04a334-26a3-497b-9db3-9bae88388209" (UID: "aa04a334-26a3-497b-9db3-9bae88388209"). InnerVolumeSpecName "kube-api-access-2lw4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.332254 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa04a334-26a3-497b-9db3-9bae88388209" (UID: "aa04a334-26a3-497b-9db3-9bae88388209"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.340467 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-config-data" (OuterVolumeSpecName: "config-data") pod "aa04a334-26a3-497b-9db3-9bae88388209" (UID: "aa04a334-26a3-497b-9db3-9bae88388209"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.350692 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "aa04a334-26a3-497b-9db3-9bae88388209" (UID: "aa04a334-26a3-497b-9db3-9bae88388209"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.396514 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.396551 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa04a334-26a3-497b-9db3-9bae88388209-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.396561 4862 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.396572 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lw4r\" (UniqueName: \"kubernetes.io/projected/aa04a334-26a3-497b-9db3-9bae88388209-kube-api-access-2lw4r\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.396583 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa04a334-26a3-497b-9db3-9bae88388209-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.636160 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.807767 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-combined-ca-bundle\") pod \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.807860 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28djv\" (UniqueName: \"kubernetes.io/projected/cbc735a4-f48b-4d86-b8f5-15fd6b417750-kube-api-access-28djv\") pod \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.807897 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-scripts\") pod \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.807916 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-sg-core-conf-yaml\") pod \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.807974 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-log-httpd\") pod \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.808024 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-run-httpd\") pod \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.808081 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-config-data\") pod \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\" (UID: \"cbc735a4-f48b-4d86-b8f5-15fd6b417750\") " Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.809016 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cbc735a4-f48b-4d86-b8f5-15fd6b417750" (UID: "cbc735a4-f48b-4d86-b8f5-15fd6b417750"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.809063 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cbc735a4-f48b-4d86-b8f5-15fd6b417750" (UID: "cbc735a4-f48b-4d86-b8f5-15fd6b417750"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.812295 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbc735a4-f48b-4d86-b8f5-15fd6b417750-kube-api-access-28djv" (OuterVolumeSpecName: "kube-api-access-28djv") pod "cbc735a4-f48b-4d86-b8f5-15fd6b417750" (UID: "cbc735a4-f48b-4d86-b8f5-15fd6b417750"). InnerVolumeSpecName "kube-api-access-28djv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.825216 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-scripts" (OuterVolumeSpecName: "scripts") pod "cbc735a4-f48b-4d86-b8f5-15fd6b417750" (UID: "cbc735a4-f48b-4d86-b8f5-15fd6b417750"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.838043 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cbc735a4-f48b-4d86-b8f5-15fd6b417750" (UID: "cbc735a4-f48b-4d86-b8f5-15fd6b417750"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.911656 4862 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.911714 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28djv\" (UniqueName: \"kubernetes.io/projected/cbc735a4-f48b-4d86-b8f5-15fd6b417750-kube-api-access-28djv\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.911727 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.911740 4862 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.911775 4862 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbc735a4-f48b-4d86-b8f5-15fd6b417750-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.913846 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-config-data" (OuterVolumeSpecName: "config-data") pod "cbc735a4-f48b-4d86-b8f5-15fd6b417750" (UID: "cbc735a4-f48b-4d86-b8f5-15fd6b417750"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:49 crc kubenswrapper[4862]: I1128 11:45:49.914776 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbc735a4-f48b-4d86-b8f5-15fd6b417750" (UID: "cbc735a4-f48b-4d86-b8f5-15fd6b417750"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.014918 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.014954 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc735a4-f48b-4d86-b8f5-15fd6b417750-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.059742 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ae02e889-f98f-4678-b15f-d91f711083f9","Type":"ContainerStarted","Data":"e487aac69ef596761ce6caee885f81d9e334146dc7a0a9ba8b24df61e5dc1e8f"} Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.060880 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.066182 4862 generic.go:334] "Generic (PLEG): container finished" podID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerID="aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830" exitCode=0 Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.066239 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.066267 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.066286 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerDied","Data":"aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830"} Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.066310 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbc735a4-f48b-4d86-b8f5-15fd6b417750","Type":"ContainerDied","Data":"01b51725cf47e644537995117e9216ae02387ad111e585cff1b1b1c2690e13db"} Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.066330 4862 scope.go:117] "RemoveContainer" containerID="93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.091253 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.09123066 podStartE2EDuration="2.09123066s" podCreationTimestamp="2025-11-28 11:45:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:50.076769803 +0000 UTC m=+1427.609283734" watchObservedRunningTime="2025-11-28 11:45:50.09123066 +0000 UTC m=+1427.623744581" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.113720 4862 scope.go:117] "RemoveContainer" containerID="070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.144040 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.158029 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.167706 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.172514 4862 scope.go:117] "RemoveContainer" containerID="aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191047 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.191464 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="sg-core" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191484 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="sg-core" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.191504 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="ceilometer-notification-agent" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191510 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="ceilometer-notification-agent" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.191531 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="ceilometer-central-agent" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191537 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="ceilometer-central-agent" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.191554 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="proxy-httpd" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191559 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="proxy-httpd" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.191579 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa04a334-26a3-497b-9db3-9bae88388209" containerName="nova-metadata-metadata" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191584 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa04a334-26a3-497b-9db3-9bae88388209" containerName="nova-metadata-metadata" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.191600 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa04a334-26a3-497b-9db3-9bae88388209" containerName="nova-metadata-log" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191605 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa04a334-26a3-497b-9db3-9bae88388209" containerName="nova-metadata-log" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191764 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="ceilometer-notification-agent" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191779 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="proxy-httpd" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191793 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="sg-core" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191804 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa04a334-26a3-497b-9db3-9bae88388209" containerName="nova-metadata-metadata" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191815 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa04a334-26a3-497b-9db3-9bae88388209" containerName="nova-metadata-log" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.191823 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" containerName="ceilometer-central-agent" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.192871 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.195216 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.195370 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.210528 4862 scope.go:117] "RemoveContainer" containerID="36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.211136 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.228436 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.235871 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.236027 4862 scope.go:117] "RemoveContainer" containerID="93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.236756 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb\": container with ID starting with 93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb not found: ID does not exist" containerID="93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.236799 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb"} err="failed to get container status \"93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb\": rpc error: code = NotFound desc = could not find container \"93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb\": container with ID starting with 93b0a6ceffed9c44afb118472412a9e892f63cc91d529a2daee6e31c185e1cfb not found: ID does not exist" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.236830 4862 scope.go:117] "RemoveContainer" containerID="070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.237286 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea\": container with ID starting with 070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea not found: ID does not exist" containerID="070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.237321 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea"} err="failed to get container status \"070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea\": rpc error: code = NotFound desc = could not find container \"070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea\": container with ID starting with 070c2f292699ef9b936a61f09fb2fb7c39209894df4237a804ed9c8e49c4a5ea not found: ID does not exist" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.237347 4862 scope.go:117] "RemoveContainer" containerID="aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.237647 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830\": container with ID starting with aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830 not found: ID does not exist" containerID="aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.237690 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830"} err="failed to get container status \"aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830\": rpc error: code = NotFound desc = could not find container \"aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830\": container with ID starting with aedef4291f9f5f4107efe20c43dae8034680927338991d0b83031d92d8715830 not found: ID does not exist" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.237713 4862 scope.go:117] "RemoveContainer" containerID="36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.238016 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6\": container with ID starting with 36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6 not found: ID does not exist" containerID="36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.238039 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6"} err="failed to get container status \"36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6\": rpc error: code = NotFound desc = could not find container \"36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6\": container with ID starting with 36a464d825d9d0e9732eb8bc95ca23a593658a81fa1137eafa040ffeec6cbcd6 not found: ID does not exist" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.238831 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.242808 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.243001 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.243204 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.244563 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.318757 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.318821 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.318869 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.318992 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.319023 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.319051 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-config-data\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.319080 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6d26\" (UniqueName: \"kubernetes.io/projected/22b82b62-3276-4527-90b3-d2abbbbb92ab-kube-api-access-c6d26\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.319207 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-log-httpd\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.319331 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b82b62-3276-4527-90b3-d2abbbbb92ab-logs\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.319512 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgvdj\" (UniqueName: \"kubernetes.io/projected/f90c0e88-8de7-41c0-a838-89df16f1024f-kube-api-access-qgvdj\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.319604 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-run-httpd\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.319660 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-config-data\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.319695 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-scripts\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.333176 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.334756 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.335992 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 11:45:50 crc kubenswrapper[4862]: E1128 11:45:50.336032 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="153e6896-1533-4559-a5ba-b4bda00fac40" containerName="nova-scheduler-scheduler" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422117 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-log-httpd\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422176 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b82b62-3276-4527-90b3-d2abbbbb92ab-logs\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422232 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgvdj\" (UniqueName: \"kubernetes.io/projected/f90c0e88-8de7-41c0-a838-89df16f1024f-kube-api-access-qgvdj\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422272 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-run-httpd\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422296 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-config-data\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422316 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-scripts\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422365 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422389 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422417 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422490 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422519 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422540 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-config-data\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422567 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6d26\" (UniqueName: \"kubernetes.io/projected/22b82b62-3276-4527-90b3-d2abbbbb92ab-kube-api-access-c6d26\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422633 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-log-httpd\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.422710 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b82b62-3276-4527-90b3-d2abbbbb92ab-logs\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.423533 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-run-httpd\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.427591 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.427936 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.428387 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-config-data\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.428919 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.430316 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.430411 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-config-data\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.431869 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.438106 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-scripts\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.439012 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgvdj\" (UniqueName: \"kubernetes.io/projected/f90c0e88-8de7-41c0-a838-89df16f1024f-kube-api-access-qgvdj\") pod \"ceilometer-0\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.449058 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6d26\" (UniqueName: \"kubernetes.io/projected/22b82b62-3276-4527-90b3-d2abbbbb92ab-kube-api-access-c6d26\") pod \"nova-metadata-0\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.512207 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.568245 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.837522 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:45:50 crc kubenswrapper[4862]: W1128 11:45:50.847885 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22b82b62_3276_4527_90b3_d2abbbbb92ab.slice/crio-99da844c8d25b38a0dfe38d51d4c8b8c1f73ebc1724f101db4a0cbb63126998f WatchSource:0}: Error finding container 99da844c8d25b38a0dfe38d51d4c8b8c1f73ebc1724f101db4a0cbb63126998f: Status 404 returned error can't find the container with id 99da844c8d25b38a0dfe38d51d4c8b8c1f73ebc1724f101db4a0cbb63126998f Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.858461 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa04a334-26a3-497b-9db3-9bae88388209" path="/var/lib/kubelet/pods/aa04a334-26a3-497b-9db3-9bae88388209/volumes" Nov 28 11:45:50 crc kubenswrapper[4862]: I1128 11:45:50.859649 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbc735a4-f48b-4d86-b8f5-15fd6b417750" path="/var/lib/kubelet/pods/cbc735a4-f48b-4d86-b8f5-15fd6b417750/volumes" Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.083391 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22b82b62-3276-4527-90b3-d2abbbbb92ab","Type":"ContainerStarted","Data":"4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0"} Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.083428 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22b82b62-3276-4527-90b3-d2abbbbb92ab","Type":"ContainerStarted","Data":"99da844c8d25b38a0dfe38d51d4c8b8c1f73ebc1724f101db4a0cbb63126998f"} Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.134669 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:45:51 crc kubenswrapper[4862]: W1128 11:45:51.140173 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf90c0e88_8de7_41c0_a838_89df16f1024f.slice/crio-05aa8e1938d95ee4b9c720281bcbe39dea898b3038e6028e1143d21b764dfcae WatchSource:0}: Error finding container 05aa8e1938d95ee4b9c720281bcbe39dea898b3038e6028e1143d21b764dfcae: Status 404 returned error can't find the container with id 05aa8e1938d95ee4b9c720281bcbe39dea898b3038e6028e1143d21b764dfcae Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.823281 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.954043 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-combined-ca-bundle\") pod \"153e6896-1533-4559-a5ba-b4bda00fac40\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.954170 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-config-data\") pod \"153e6896-1533-4559-a5ba-b4bda00fac40\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.955026 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzw8c\" (UniqueName: \"kubernetes.io/projected/153e6896-1533-4559-a5ba-b4bda00fac40-kube-api-access-zzw8c\") pod \"153e6896-1533-4559-a5ba-b4bda00fac40\" (UID: \"153e6896-1533-4559-a5ba-b4bda00fac40\") " Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.962493 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/153e6896-1533-4559-a5ba-b4bda00fac40-kube-api-access-zzw8c" (OuterVolumeSpecName: "kube-api-access-zzw8c") pod "153e6896-1533-4559-a5ba-b4bda00fac40" (UID: "153e6896-1533-4559-a5ba-b4bda00fac40"). InnerVolumeSpecName "kube-api-access-zzw8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.983807 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-config-data" (OuterVolumeSpecName: "config-data") pod "153e6896-1533-4559-a5ba-b4bda00fac40" (UID: "153e6896-1533-4559-a5ba-b4bda00fac40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:51 crc kubenswrapper[4862]: I1128 11:45:51.992527 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "153e6896-1533-4559-a5ba-b4bda00fac40" (UID: "153e6896-1533-4559-a5ba-b4bda00fac40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.058662 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzw8c\" (UniqueName: \"kubernetes.io/projected/153e6896-1533-4559-a5ba-b4bda00fac40-kube-api-access-zzw8c\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.058694 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.058704 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153e6896-1533-4559-a5ba-b4bda00fac40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.091742 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerStarted","Data":"700a6ae64c6f634ec8de8f0e56d2e9e04f29aaf0eed950337541b42bcae36a57"} Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.091782 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerStarted","Data":"05aa8e1938d95ee4b9c720281bcbe39dea898b3038e6028e1143d21b764dfcae"} Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.093463 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22b82b62-3276-4527-90b3-d2abbbbb92ab","Type":"ContainerStarted","Data":"a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49"} Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.095877 4862 generic.go:334] "Generic (PLEG): container finished" podID="153e6896-1533-4559-a5ba-b4bda00fac40" containerID="a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276" exitCode=0 Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.096402 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.103177 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"153e6896-1533-4559-a5ba-b4bda00fac40","Type":"ContainerDied","Data":"a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276"} Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.103238 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"153e6896-1533-4559-a5ba-b4bda00fac40","Type":"ContainerDied","Data":"de4bb2b9c942cda17adfeb34b6132e48aca782ecfe832881a1f91e3778bce288"} Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.103256 4862 scope.go:117] "RemoveContainer" containerID="a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.128617 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.128597862 podStartE2EDuration="2.128597862s" podCreationTimestamp="2025-11-28 11:45:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:52.122335282 +0000 UTC m=+1429.654849203" watchObservedRunningTime="2025-11-28 11:45:52.128597862 +0000 UTC m=+1429.661111783" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.135896 4862 scope.go:117] "RemoveContainer" containerID="a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276" Nov 28 11:45:52 crc kubenswrapper[4862]: E1128 11:45:52.136395 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276\": container with ID starting with a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276 not found: ID does not exist" containerID="a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.136425 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276"} err="failed to get container status \"a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276\": rpc error: code = NotFound desc = could not find container \"a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276\": container with ID starting with a4d277c72a5cb4d0f1750549c14133134a3445be32d0621ae46a0b7229d0d276 not found: ID does not exist" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.156718 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.164566 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.179258 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:45:52 crc kubenswrapper[4862]: E1128 11:45:52.179863 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="153e6896-1533-4559-a5ba-b4bda00fac40" containerName="nova-scheduler-scheduler" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.179887 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="153e6896-1533-4559-a5ba-b4bda00fac40" containerName="nova-scheduler-scheduler" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.180142 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="153e6896-1533-4559-a5ba-b4bda00fac40" containerName="nova-scheduler-scheduler" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.181066 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.184228 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.195396 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.262658 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-config-data\") pod \"nova-scheduler-0\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.263658 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.263733 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8fpg\" (UniqueName: \"kubernetes.io/projected/b1edabad-9cf9-4e4f-935a-01493ebce5ec-kube-api-access-h8fpg\") pod \"nova-scheduler-0\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.365192 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-config-data\") pod \"nova-scheduler-0\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.365290 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.365365 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8fpg\" (UniqueName: \"kubernetes.io/projected/b1edabad-9cf9-4e4f-935a-01493ebce5ec-kube-api-access-h8fpg\") pod \"nova-scheduler-0\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.370620 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-config-data\") pod \"nova-scheduler-0\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.370886 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.381834 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8fpg\" (UniqueName: \"kubernetes.io/projected/b1edabad-9cf9-4e4f-935a-01493ebce5ec-kube-api-access-h8fpg\") pod \"nova-scheduler-0\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.502325 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.871878 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="153e6896-1533-4559-a5ba-b4bda00fac40" path="/var/lib/kubelet/pods/153e6896-1533-4559-a5ba-b4bda00fac40/volumes" Nov 28 11:45:52 crc kubenswrapper[4862]: I1128 11:45:52.965644 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.029133 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.108630 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1edabad-9cf9-4e4f-935a-01493ebce5ec","Type":"ContainerStarted","Data":"7c5f846f1d287be1a630204bfc75fc1714b01a082519ce675de2d59fada36438"} Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.128469 4862 generic.go:334] "Generic (PLEG): container finished" podID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerID="af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b" exitCode=0 Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.128525 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"678dbe9d-368e-48c2-9a21-42a18b2efeba","Type":"ContainerDied","Data":"af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b"} Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.128585 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"678dbe9d-368e-48c2-9a21-42a18b2efeba","Type":"ContainerDied","Data":"6113fffa459855f6bcf5bd0866d75172f1ec090299c4a8cf67ac8de8c737cf38"} Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.128544 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.128603 4862 scope.go:117] "RemoveContainer" containerID="af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.150110 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerStarted","Data":"b6daeab2cb9e0eab83885f7c09a724e82f74ac5e24d33148a1716002cfa0c5fb"} Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.153436 4862 scope.go:117] "RemoveContainer" containerID="46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.170014 4862 scope.go:117] "RemoveContainer" containerID="af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b" Nov 28 11:45:53 crc kubenswrapper[4862]: E1128 11:45:53.170468 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b\": container with ID starting with af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b not found: ID does not exist" containerID="af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.170495 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b"} err="failed to get container status \"af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b\": rpc error: code = NotFound desc = could not find container \"af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b\": container with ID starting with af03ac43bfba048dbb3d4d9fc9c42b988dba7f8e717621cc2602b618de2cd10b not found: ID does not exist" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.170513 4862 scope.go:117] "RemoveContainer" containerID="46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778" Nov 28 11:45:53 crc kubenswrapper[4862]: E1128 11:45:53.170781 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778\": container with ID starting with 46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778 not found: ID does not exist" containerID="46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.170804 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778"} err="failed to get container status \"46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778\": rpc error: code = NotFound desc = could not find container \"46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778\": container with ID starting with 46f6340d8d4c597f1e188bb61789c283e903a15b54e22012f14e490253a8e778 not found: ID does not exist" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.182494 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrm84\" (UniqueName: \"kubernetes.io/projected/678dbe9d-368e-48c2-9a21-42a18b2efeba-kube-api-access-zrm84\") pod \"678dbe9d-368e-48c2-9a21-42a18b2efeba\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.182599 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-config-data\") pod \"678dbe9d-368e-48c2-9a21-42a18b2efeba\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.182672 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-combined-ca-bundle\") pod \"678dbe9d-368e-48c2-9a21-42a18b2efeba\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.182773 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/678dbe9d-368e-48c2-9a21-42a18b2efeba-logs\") pod \"678dbe9d-368e-48c2-9a21-42a18b2efeba\" (UID: \"678dbe9d-368e-48c2-9a21-42a18b2efeba\") " Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.183804 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/678dbe9d-368e-48c2-9a21-42a18b2efeba-logs" (OuterVolumeSpecName: "logs") pod "678dbe9d-368e-48c2-9a21-42a18b2efeba" (UID: "678dbe9d-368e-48c2-9a21-42a18b2efeba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.187634 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/678dbe9d-368e-48c2-9a21-42a18b2efeba-kube-api-access-zrm84" (OuterVolumeSpecName: "kube-api-access-zrm84") pod "678dbe9d-368e-48c2-9a21-42a18b2efeba" (UID: "678dbe9d-368e-48c2-9a21-42a18b2efeba"). InnerVolumeSpecName "kube-api-access-zrm84". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.215028 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "678dbe9d-368e-48c2-9a21-42a18b2efeba" (UID: "678dbe9d-368e-48c2-9a21-42a18b2efeba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.215184 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-config-data" (OuterVolumeSpecName: "config-data") pod "678dbe9d-368e-48c2-9a21-42a18b2efeba" (UID: "678dbe9d-368e-48c2-9a21-42a18b2efeba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.285559 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/678dbe9d-368e-48c2-9a21-42a18b2efeba-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.285590 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrm84\" (UniqueName: \"kubernetes.io/projected/678dbe9d-368e-48c2-9a21-42a18b2efeba-kube-api-access-zrm84\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.285601 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.285610 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678dbe9d-368e-48c2-9a21-42a18b2efeba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.467248 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.476616 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.501181 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 11:45:53 crc kubenswrapper[4862]: E1128 11:45:53.501640 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-api" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.501660 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-api" Nov 28 11:45:53 crc kubenswrapper[4862]: E1128 11:45:53.501677 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-log" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.501684 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-log" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.501860 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-log" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.501882 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" containerName="nova-api-api" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.502869 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.505415 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.508854 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.591451 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc4m5\" (UniqueName: \"kubernetes.io/projected/e8c408c5-bcf9-4a70-8954-016ec8618502-kube-api-access-vc4m5\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.591506 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.591627 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8c408c5-bcf9-4a70-8954-016ec8618502-logs\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.591778 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-config-data\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.693880 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc4m5\" (UniqueName: \"kubernetes.io/projected/e8c408c5-bcf9-4a70-8954-016ec8618502-kube-api-access-vc4m5\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.694043 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.694276 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8c408c5-bcf9-4a70-8954-016ec8618502-logs\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.694392 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-config-data\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.694786 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8c408c5-bcf9-4a70-8954-016ec8618502-logs\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.700756 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.700944 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-config-data\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.719583 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc4m5\" (UniqueName: \"kubernetes.io/projected/e8c408c5-bcf9-4a70-8954-016ec8618502-kube-api-access-vc4m5\") pod \"nova-api-0\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " pod="openstack/nova-api-0" Nov 28 11:45:53 crc kubenswrapper[4862]: I1128 11:45:53.853757 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:45:54 crc kubenswrapper[4862]: I1128 11:45:54.165964 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerStarted","Data":"a09438ad6c8003f94013130e6fbae489459ae08e8d3c237dd5a9dfd08caebdfd"} Nov 28 11:45:54 crc kubenswrapper[4862]: I1128 11:45:54.169766 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1edabad-9cf9-4e4f-935a-01493ebce5ec","Type":"ContainerStarted","Data":"921db04cc1e5a5fe187dff1e200d67d69b50f9c36daae90f7a96a302f3ad65cc"} Nov 28 11:45:54 crc kubenswrapper[4862]: I1128 11:45:54.193019 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.193004143 podStartE2EDuration="2.193004143s" podCreationTimestamp="2025-11-28 11:45:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:54.18968042 +0000 UTC m=+1431.722194341" watchObservedRunningTime="2025-11-28 11:45:54.193004143 +0000 UTC m=+1431.725518064" Nov 28 11:45:54 crc kubenswrapper[4862]: I1128 11:45:54.304167 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:45:54 crc kubenswrapper[4862]: W1128 11:45:54.316382 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8c408c5_bcf9_4a70_8954_016ec8618502.slice/crio-0b3863aa329a618fcd20ac298dc8c4e49f8e02c9b586ff7a641b78d9086e6230 WatchSource:0}: Error finding container 0b3863aa329a618fcd20ac298dc8c4e49f8e02c9b586ff7a641b78d9086e6230: Status 404 returned error can't find the container with id 0b3863aa329a618fcd20ac298dc8c4e49f8e02c9b586ff7a641b78d9086e6230 Nov 28 11:45:54 crc kubenswrapper[4862]: I1128 11:45:54.855738 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="678dbe9d-368e-48c2-9a21-42a18b2efeba" path="/var/lib/kubelet/pods/678dbe9d-368e-48c2-9a21-42a18b2efeba/volumes" Nov 28 11:45:55 crc kubenswrapper[4862]: I1128 11:45:55.183236 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerStarted","Data":"4e3b49e4b2deac8f52253fb5568368b7416015ed50d5db8ec5c75aab2135ac13"} Nov 28 11:45:55 crc kubenswrapper[4862]: I1128 11:45:55.183624 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 11:45:55 crc kubenswrapper[4862]: I1128 11:45:55.185420 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e8c408c5-bcf9-4a70-8954-016ec8618502","Type":"ContainerStarted","Data":"cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00"} Nov 28 11:45:55 crc kubenswrapper[4862]: I1128 11:45:55.185473 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e8c408c5-bcf9-4a70-8954-016ec8618502","Type":"ContainerStarted","Data":"cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09"} Nov 28 11:45:55 crc kubenswrapper[4862]: I1128 11:45:55.185491 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e8c408c5-bcf9-4a70-8954-016ec8618502","Type":"ContainerStarted","Data":"0b3863aa329a618fcd20ac298dc8c4e49f8e02c9b586ff7a641b78d9086e6230"} Nov 28 11:45:55 crc kubenswrapper[4862]: I1128 11:45:55.205953 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.425449651 podStartE2EDuration="5.205933829s" podCreationTimestamp="2025-11-28 11:45:50 +0000 UTC" firstStartedPulling="2025-11-28 11:45:51.142646733 +0000 UTC m=+1428.675160654" lastFinishedPulling="2025-11-28 11:45:54.923130881 +0000 UTC m=+1432.455644832" observedRunningTime="2025-11-28 11:45:55.202071709 +0000 UTC m=+1432.734585650" watchObservedRunningTime="2025-11-28 11:45:55.205933829 +0000 UTC m=+1432.738447750" Nov 28 11:45:55 crc kubenswrapper[4862]: I1128 11:45:55.220980 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.220960903 podStartE2EDuration="2.220960903s" podCreationTimestamp="2025-11-28 11:45:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:45:55.219421033 +0000 UTC m=+1432.751934954" watchObservedRunningTime="2025-11-28 11:45:55.220960903 +0000 UTC m=+1432.753474824" Nov 28 11:45:55 crc kubenswrapper[4862]: I1128 11:45:55.512957 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 11:45:55 crc kubenswrapper[4862]: I1128 11:45:55.513009 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 11:45:56 crc kubenswrapper[4862]: I1128 11:45:56.382231 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 28 11:45:57 crc kubenswrapper[4862]: I1128 11:45:57.503144 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 11:45:58 crc kubenswrapper[4862]: I1128 11:45:58.462686 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 28 11:46:00 crc kubenswrapper[4862]: I1128 11:46:00.512762 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 11:46:00 crc kubenswrapper[4862]: I1128 11:46:00.513019 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 11:46:01 crc kubenswrapper[4862]: I1128 11:46:01.531338 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:01 crc kubenswrapper[4862]: I1128 11:46:01.531346 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:02 crc kubenswrapper[4862]: I1128 11:46:02.503149 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 11:46:02 crc kubenswrapper[4862]: I1128 11:46:02.555855 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 11:46:03 crc kubenswrapper[4862]: I1128 11:46:03.378549 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 11:46:03 crc kubenswrapper[4862]: I1128 11:46:03.854266 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 11:46:03 crc kubenswrapper[4862]: I1128 11:46:03.854318 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 11:46:04 crc kubenswrapper[4862]: I1128 11:46:04.899277 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:04 crc kubenswrapper[4862]: I1128 11:46:04.940563 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:08 crc kubenswrapper[4862]: I1128 11:46:08.291630 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:46:08 crc kubenswrapper[4862]: I1128 11:46:08.293732 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:46:08 crc kubenswrapper[4862]: I1128 11:46:08.293965 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:46:08 crc kubenswrapper[4862]: I1128 11:46:08.295303 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f1769f4e9e7625f7b202c55e00f4690f331fdd2c1054c72cfd81ec88107ff67"} pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:46:08 crc kubenswrapper[4862]: I1128 11:46:08.295580 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" containerID="cri-o://9f1769f4e9e7625f7b202c55e00f4690f331fdd2c1054c72cfd81ec88107ff67" gracePeriod=600 Nov 28 11:46:09 crc kubenswrapper[4862]: I1128 11:46:09.361575 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6845439-1a4c-4f28-9ba4-29149d545995" containerID="9f1769f4e9e7625f7b202c55e00f4690f331fdd2c1054c72cfd81ec88107ff67" exitCode=0 Nov 28 11:46:09 crc kubenswrapper[4862]: I1128 11:46:09.361634 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerDied","Data":"9f1769f4e9e7625f7b202c55e00f4690f331fdd2c1054c72cfd81ec88107ff67"} Nov 28 11:46:09 crc kubenswrapper[4862]: I1128 11:46:09.362126 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776"} Nov 28 11:46:09 crc kubenswrapper[4862]: I1128 11:46:09.362156 4862 scope.go:117] "RemoveContainer" containerID="1439dec9a32df00b0169c77b7376186b3da92973a1be5f562a161099f8fb96c8" Nov 28 11:46:10 crc kubenswrapper[4862]: I1128 11:46:10.519387 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 11:46:10 crc kubenswrapper[4862]: I1128 11:46:10.523620 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 11:46:10 crc kubenswrapper[4862]: I1128 11:46:10.524920 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.285143 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.296905 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-config-data\") pod \"bad61764-bb4e-4bab-9712-e1b83597b642\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.297131 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm7tw\" (UniqueName: \"kubernetes.io/projected/bad61764-bb4e-4bab-9712-e1b83597b642-kube-api-access-gm7tw\") pod \"bad61764-bb4e-4bab-9712-e1b83597b642\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.297228 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-combined-ca-bundle\") pod \"bad61764-bb4e-4bab-9712-e1b83597b642\" (UID: \"bad61764-bb4e-4bab-9712-e1b83597b642\") " Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.305948 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad61764-bb4e-4bab-9712-e1b83597b642-kube-api-access-gm7tw" (OuterVolumeSpecName: "kube-api-access-gm7tw") pod "bad61764-bb4e-4bab-9712-e1b83597b642" (UID: "bad61764-bb4e-4bab-9712-e1b83597b642"). InnerVolumeSpecName "kube-api-access-gm7tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.341101 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bad61764-bb4e-4bab-9712-e1b83597b642" (UID: "bad61764-bb4e-4bab-9712-e1b83597b642"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.354289 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-config-data" (OuterVolumeSpecName: "config-data") pod "bad61764-bb4e-4bab-9712-e1b83597b642" (UID: "bad61764-bb4e-4bab-9712-e1b83597b642"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.388705 4862 generic.go:334] "Generic (PLEG): container finished" podID="bad61764-bb4e-4bab-9712-e1b83597b642" containerID="dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca" exitCode=137 Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.388815 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.388821 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bad61764-bb4e-4bab-9712-e1b83597b642","Type":"ContainerDied","Data":"dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca"} Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.389133 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bad61764-bb4e-4bab-9712-e1b83597b642","Type":"ContainerDied","Data":"66c168311bc5f917103b0d6666120c76d65a9928a3f63a0b409df23adec739d8"} Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.389174 4862 scope.go:117] "RemoveContainer" containerID="dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.401389 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm7tw\" (UniqueName: \"kubernetes.io/projected/bad61764-bb4e-4bab-9712-e1b83597b642-kube-api-access-gm7tw\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.401699 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.401857 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad61764-bb4e-4bab-9712-e1b83597b642-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.420847 4862 scope.go:117] "RemoveContainer" containerID="dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca" Nov 28 11:46:11 crc kubenswrapper[4862]: E1128 11:46:11.421549 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca\": container with ID starting with dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca not found: ID does not exist" containerID="dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.421610 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca"} err="failed to get container status \"dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca\": rpc error: code = NotFound desc = could not find container \"dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca\": container with ID starting with dfb629221db5e9d38aaddab63fc6d3f8b055dbaacff6ee3686e32b0fc0bc0aca not found: ID does not exist" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.423744 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.451655 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.474223 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.493004 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:46:11 crc kubenswrapper[4862]: E1128 11:46:11.493823 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad61764-bb4e-4bab-9712-e1b83597b642" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.493854 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad61764-bb4e-4bab-9712-e1b83597b642" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.494204 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad61764-bb4e-4bab-9712-e1b83597b642" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.495369 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.499271 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.506125 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.506907 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.507526 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.607835 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.607905 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtxql\" (UniqueName: \"kubernetes.io/projected/73c57ace-65f0-40ab-996a-e5595851177a-kube-api-access-xtxql\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.609283 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.609438 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.609511 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.703529 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wrwbg"] Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.705453 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.715370 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wrwbg"] Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.715869 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-utilities\") pod \"redhat-operators-wrwbg\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.715954 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.715988 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-catalog-content\") pod \"redhat-operators-wrwbg\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.716009 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtxql\" (UniqueName: \"kubernetes.io/projected/73c57ace-65f0-40ab-996a-e5595851177a-kube-api-access-xtxql\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.716041 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.716071 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.716110 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.716130 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnczv\" (UniqueName: \"kubernetes.io/projected/4b3b2737-87e3-4017-965e-3a49a88c4526-kube-api-access-dnczv\") pod \"redhat-operators-wrwbg\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.720561 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.724581 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.725115 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.731677 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.751912 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtxql\" (UniqueName: \"kubernetes.io/projected/73c57ace-65f0-40ab-996a-e5595851177a-kube-api-access-xtxql\") pod \"nova-cell1-novncproxy-0\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.817931 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-utilities\") pod \"redhat-operators-wrwbg\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.818062 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-catalog-content\") pod \"redhat-operators-wrwbg\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.818155 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnczv\" (UniqueName: \"kubernetes.io/projected/4b3b2737-87e3-4017-965e-3a49a88c4526-kube-api-access-dnczv\") pod \"redhat-operators-wrwbg\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.818487 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-utilities\") pod \"redhat-operators-wrwbg\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.818624 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-catalog-content\") pod \"redhat-operators-wrwbg\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.838907 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnczv\" (UniqueName: \"kubernetes.io/projected/4b3b2737-87e3-4017-965e-3a49a88c4526-kube-api-access-dnczv\") pod \"redhat-operators-wrwbg\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:11 crc kubenswrapper[4862]: I1128 11:46:11.840839 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:12 crc kubenswrapper[4862]: I1128 11:46:12.036059 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:12 crc kubenswrapper[4862]: I1128 11:46:12.173620 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:46:12 crc kubenswrapper[4862]: I1128 11:46:12.418755 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"73c57ace-65f0-40ab-996a-e5595851177a","Type":"ContainerStarted","Data":"f47d98555cb4c9a93ed28f77805a70993ecebef110883d750502c0b7280310a3"} Nov 28 11:46:12 crc kubenswrapper[4862]: I1128 11:46:12.625753 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wrwbg"] Nov 28 11:46:12 crc kubenswrapper[4862]: I1128 11:46:12.848223 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bad61764-bb4e-4bab-9712-e1b83597b642" path="/var/lib/kubelet/pods/bad61764-bb4e-4bab-9712-e1b83597b642/volumes" Nov 28 11:46:13 crc kubenswrapper[4862]: I1128 11:46:13.427388 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"73c57ace-65f0-40ab-996a-e5595851177a","Type":"ContainerStarted","Data":"e4fbe66671af80c3245a6da3f4c0fcf1f2a7319780d1ab11a9026ea78828037b"} Nov 28 11:46:13 crc kubenswrapper[4862]: I1128 11:46:13.436182 4862 generic.go:334] "Generic (PLEG): container finished" podID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerID="597bcb30c2ec13a42a442bccc8d17ef41baa8183434146aff53ffd1727940883" exitCode=0 Nov 28 11:46:13 crc kubenswrapper[4862]: I1128 11:46:13.436289 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrwbg" event={"ID":"4b3b2737-87e3-4017-965e-3a49a88c4526","Type":"ContainerDied","Data":"597bcb30c2ec13a42a442bccc8d17ef41baa8183434146aff53ffd1727940883"} Nov 28 11:46:13 crc kubenswrapper[4862]: I1128 11:46:13.436365 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrwbg" event={"ID":"4b3b2737-87e3-4017-965e-3a49a88c4526","Type":"ContainerStarted","Data":"51af13f1d7e96a7dd7217b53a3143b46046d278bb49ce435bb883d7060d5d5cb"} Nov 28 11:46:13 crc kubenswrapper[4862]: I1128 11:46:13.461419 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.461401758 podStartE2EDuration="2.461401758s" podCreationTimestamp="2025-11-28 11:46:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:46:13.441735514 +0000 UTC m=+1450.974249435" watchObservedRunningTime="2025-11-28 11:46:13.461401758 +0000 UTC m=+1450.993915679" Nov 28 11:46:13 crc kubenswrapper[4862]: I1128 11:46:13.867174 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 11:46:13 crc kubenswrapper[4862]: I1128 11:46:13.868023 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 11:46:13 crc kubenswrapper[4862]: I1128 11:46:13.868677 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 11:46:13 crc kubenswrapper[4862]: I1128 11:46:13.871957 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.457083 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrwbg" event={"ID":"4b3b2737-87e3-4017-965e-3a49a88c4526","Type":"ContainerStarted","Data":"f1560a3f2403b0cebf9942cbcd616447fa7c77edc5218f4969d1d49654b4431d"} Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.457497 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.465770 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.688705 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc646c8f9-jmhdw"] Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.695617 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.717493 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc646c8f9-jmhdw"] Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.881242 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-nb\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.881425 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw9x6\" (UniqueName: \"kubernetes.io/projected/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-kube-api-access-kw9x6\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.881681 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-swift-storage-0\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.881742 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.881912 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-config\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.881982 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-svc\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.983669 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw9x6\" (UniqueName: \"kubernetes.io/projected/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-kube-api-access-kw9x6\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.983779 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-swift-storage-0\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.983812 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.983905 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-config\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.983943 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-svc\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.984062 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-nb\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.985061 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-nb\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.985379 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-config\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.985546 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.985807 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-svc\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:14 crc kubenswrapper[4862]: I1128 11:46:14.987547 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-swift-storage-0\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:15 crc kubenswrapper[4862]: I1128 11:46:15.004384 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw9x6\" (UniqueName: \"kubernetes.io/projected/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-kube-api-access-kw9x6\") pod \"dnsmasq-dns-6bc646c8f9-jmhdw\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:15 crc kubenswrapper[4862]: I1128 11:46:15.033227 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:15 crc kubenswrapper[4862]: I1128 11:46:15.453767 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc646c8f9-jmhdw"] Nov 28 11:46:15 crc kubenswrapper[4862]: I1128 11:46:15.494829 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" event={"ID":"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5","Type":"ContainerStarted","Data":"c83ad775ae3fa88812e0b5fcde9332a702ac484f58ea3c05ae2742a8b43e2e66"} Nov 28 11:46:16 crc kubenswrapper[4862]: I1128 11:46:16.517103 4862 generic.go:334] "Generic (PLEG): container finished" podID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerID="f1560a3f2403b0cebf9942cbcd616447fa7c77edc5218f4969d1d49654b4431d" exitCode=0 Nov 28 11:46:16 crc kubenswrapper[4862]: I1128 11:46:16.517227 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrwbg" event={"ID":"4b3b2737-87e3-4017-965e-3a49a88c4526","Type":"ContainerDied","Data":"f1560a3f2403b0cebf9942cbcd616447fa7c77edc5218f4969d1d49654b4431d"} Nov 28 11:46:16 crc kubenswrapper[4862]: I1128 11:46:16.520749 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" event={"ID":"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5","Type":"ContainerStarted","Data":"f3dedbee71dddb90736b530eb1e7f1b62741d89591234e177e6f9c26bdde78a3"} Nov 28 11:46:16 crc kubenswrapper[4862]: I1128 11:46:16.850259 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:17 crc kubenswrapper[4862]: I1128 11:46:17.541268 4862 generic.go:334] "Generic (PLEG): container finished" podID="106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" containerID="f3dedbee71dddb90736b530eb1e7f1b62741d89591234e177e6f9c26bdde78a3" exitCode=0 Nov 28 11:46:17 crc kubenswrapper[4862]: I1128 11:46:17.541322 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" event={"ID":"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5","Type":"ContainerDied","Data":"f3dedbee71dddb90736b530eb1e7f1b62741d89591234e177e6f9c26bdde78a3"} Nov 28 11:46:17 crc kubenswrapper[4862]: I1128 11:46:17.599444 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:17 crc kubenswrapper[4862]: I1128 11:46:17.599965 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-log" containerID="cri-o://cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09" gracePeriod=30 Nov 28 11:46:17 crc kubenswrapper[4862]: I1128 11:46:17.600373 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-api" containerID="cri-o://cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00" gracePeriod=30 Nov 28 11:46:18 crc kubenswrapper[4862]: I1128 11:46:18.553326 4862 generic.go:334] "Generic (PLEG): container finished" podID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerID="cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09" exitCode=143 Nov 28 11:46:18 crc kubenswrapper[4862]: I1128 11:46:18.553414 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e8c408c5-bcf9-4a70-8954-016ec8618502","Type":"ContainerDied","Data":"cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09"} Nov 28 11:46:18 crc kubenswrapper[4862]: I1128 11:46:18.556177 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" event={"ID":"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5","Type":"ContainerStarted","Data":"d351e93457fe5a70f380ffce1f54e4bea632ff6f5419983422bc2fc2947f982d"} Nov 28 11:46:18 crc kubenswrapper[4862]: I1128 11:46:18.556261 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:18 crc kubenswrapper[4862]: I1128 11:46:18.561623 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrwbg" event={"ID":"4b3b2737-87e3-4017-965e-3a49a88c4526","Type":"ContainerStarted","Data":"bafdf6e87c9e4c8287da1ee020aafbb27a9b6879dd62516aa2058138e4476799"} Nov 28 11:46:18 crc kubenswrapper[4862]: I1128 11:46:18.583463 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" podStartSLOduration=4.583439491 podStartE2EDuration="4.583439491s" podCreationTimestamp="2025-11-28 11:46:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:46:18.579246017 +0000 UTC m=+1456.111759948" watchObservedRunningTime="2025-11-28 11:46:18.583439491 +0000 UTC m=+1456.115953412" Nov 28 11:46:18 crc kubenswrapper[4862]: I1128 11:46:18.602886 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wrwbg" podStartSLOduration=3.193817867 podStartE2EDuration="7.602866322s" podCreationTimestamp="2025-11-28 11:46:11 +0000 UTC" firstStartedPulling="2025-11-28 11:46:13.439583687 +0000 UTC m=+1450.972097608" lastFinishedPulling="2025-11-28 11:46:17.848632142 +0000 UTC m=+1455.381146063" observedRunningTime="2025-11-28 11:46:18.599343096 +0000 UTC m=+1456.131857017" watchObservedRunningTime="2025-11-28 11:46:18.602866322 +0000 UTC m=+1456.135380243" Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.090295 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.090549 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="ceilometer-central-agent" containerID="cri-o://700a6ae64c6f634ec8de8f0e56d2e9e04f29aaf0eed950337541b42bcae36a57" gracePeriod=30 Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.090801 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="ceilometer-notification-agent" containerID="cri-o://b6daeab2cb9e0eab83885f7c09a724e82f74ac5e24d33148a1716002cfa0c5fb" gracePeriod=30 Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.090816 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="sg-core" containerID="cri-o://a09438ad6c8003f94013130e6fbae489459ae08e8d3c237dd5a9dfd08caebdfd" gracePeriod=30 Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.090775 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="proxy-httpd" containerID="cri-o://4e3b49e4b2deac8f52253fb5568368b7416015ed50d5db8ec5c75aab2135ac13" gracePeriod=30 Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.102904 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.191:3000/\": read tcp 10.217.0.2:50756->10.217.0.191:3000: read: connection reset by peer" Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.581969 4862 generic.go:334] "Generic (PLEG): container finished" podID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerID="4e3b49e4b2deac8f52253fb5568368b7416015ed50d5db8ec5c75aab2135ac13" exitCode=0 Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.582301 4862 generic.go:334] "Generic (PLEG): container finished" podID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerID="a09438ad6c8003f94013130e6fbae489459ae08e8d3c237dd5a9dfd08caebdfd" exitCode=2 Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.582310 4862 generic.go:334] "Generic (PLEG): container finished" podID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerID="700a6ae64c6f634ec8de8f0e56d2e9e04f29aaf0eed950337541b42bcae36a57" exitCode=0 Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.582050 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerDied","Data":"4e3b49e4b2deac8f52253fb5568368b7416015ed50d5db8ec5c75aab2135ac13"} Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.582342 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerDied","Data":"a09438ad6c8003f94013130e6fbae489459ae08e8d3c237dd5a9dfd08caebdfd"} Nov 28 11:46:19 crc kubenswrapper[4862]: I1128 11:46:19.582354 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerDied","Data":"700a6ae64c6f634ec8de8f0e56d2e9e04f29aaf0eed950337541b42bcae36a57"} Nov 28 11:46:20 crc kubenswrapper[4862]: I1128 11:46:20.570268 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.191:3000/\": dial tcp 10.217.0.191:3000: connect: connection refused" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.238401 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.322728 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-config-data\") pod \"e8c408c5-bcf9-4a70-8954-016ec8618502\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.322869 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vc4m5\" (UniqueName: \"kubernetes.io/projected/e8c408c5-bcf9-4a70-8954-016ec8618502-kube-api-access-vc4m5\") pod \"e8c408c5-bcf9-4a70-8954-016ec8618502\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.322896 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-combined-ca-bundle\") pod \"e8c408c5-bcf9-4a70-8954-016ec8618502\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.322955 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8c408c5-bcf9-4a70-8954-016ec8618502-logs\") pod \"e8c408c5-bcf9-4a70-8954-016ec8618502\" (UID: \"e8c408c5-bcf9-4a70-8954-016ec8618502\") " Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.323446 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8c408c5-bcf9-4a70-8954-016ec8618502-logs" (OuterVolumeSpecName: "logs") pod "e8c408c5-bcf9-4a70-8954-016ec8618502" (UID: "e8c408c5-bcf9-4a70-8954-016ec8618502"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.330530 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8c408c5-bcf9-4a70-8954-016ec8618502-kube-api-access-vc4m5" (OuterVolumeSpecName: "kube-api-access-vc4m5") pod "e8c408c5-bcf9-4a70-8954-016ec8618502" (UID: "e8c408c5-bcf9-4a70-8954-016ec8618502"). InnerVolumeSpecName "kube-api-access-vc4m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.354487 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8c408c5-bcf9-4a70-8954-016ec8618502" (UID: "e8c408c5-bcf9-4a70-8954-016ec8618502"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.359376 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-config-data" (OuterVolumeSpecName: "config-data") pod "e8c408c5-bcf9-4a70-8954-016ec8618502" (UID: "e8c408c5-bcf9-4a70-8954-016ec8618502"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.428469 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.428518 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vc4m5\" (UniqueName: \"kubernetes.io/projected/e8c408c5-bcf9-4a70-8954-016ec8618502-kube-api-access-vc4m5\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.428532 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c408c5-bcf9-4a70-8954-016ec8618502-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.428544 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8c408c5-bcf9-4a70-8954-016ec8618502-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.606049 4862 generic.go:334] "Generic (PLEG): container finished" podID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerID="cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00" exitCode=0 Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.606133 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e8c408c5-bcf9-4a70-8954-016ec8618502","Type":"ContainerDied","Data":"cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00"} Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.606200 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e8c408c5-bcf9-4a70-8954-016ec8618502","Type":"ContainerDied","Data":"0b3863aa329a618fcd20ac298dc8c4e49f8e02c9b586ff7a641b78d9086e6230"} Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.606220 4862 scope.go:117] "RemoveContainer" containerID="cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.606241 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.635955 4862 scope.go:117] "RemoveContainer" containerID="cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.656433 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.668314 4862 scope.go:117] "RemoveContainer" containerID="cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00" Nov 28 11:46:21 crc kubenswrapper[4862]: E1128 11:46:21.668823 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00\": container with ID starting with cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00 not found: ID does not exist" containerID="cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.668886 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00"} err="failed to get container status \"cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00\": rpc error: code = NotFound desc = could not find container \"cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00\": container with ID starting with cdce4ace765be35800d4f10ea961a48f450445ebcdf097eaba27b3b66ca70c00 not found: ID does not exist" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.668926 4862 scope.go:117] "RemoveContainer" containerID="cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09" Nov 28 11:46:21 crc kubenswrapper[4862]: E1128 11:46:21.669411 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09\": container with ID starting with cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09 not found: ID does not exist" containerID="cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.669449 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09"} err="failed to get container status \"cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09\": rpc error: code = NotFound desc = could not find container \"cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09\": container with ID starting with cfdeeb316878645e0ec991b4026d7945168de736526c9b38576810873197ad09 not found: ID does not exist" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.675255 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.694776 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:21 crc kubenswrapper[4862]: E1128 11:46:21.695275 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-log" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.695294 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-log" Nov 28 11:46:21 crc kubenswrapper[4862]: E1128 11:46:21.695313 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-api" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.695319 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-api" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.695505 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-api" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.695536 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" containerName="nova-api-log" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.696540 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.698940 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.699141 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.699445 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.706575 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.835071 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.835436 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtlkq\" (UniqueName: \"kubernetes.io/projected/4a26f616-4cf2-441b-b61f-606edcb36a55-kube-api-access-dtlkq\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.835464 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.835716 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-config-data\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.835811 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-public-tls-certs\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.835872 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a26f616-4cf2-441b-b61f-606edcb36a55-logs\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.841384 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.864495 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.938056 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.938104 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtlkq\" (UniqueName: \"kubernetes.io/projected/4a26f616-4cf2-441b-b61f-606edcb36a55-kube-api-access-dtlkq\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.938132 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.938207 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-config-data\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.938228 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-public-tls-certs\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.938248 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a26f616-4cf2-441b-b61f-606edcb36a55-logs\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.939293 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a26f616-4cf2-441b-b61f-606edcb36a55-logs\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.943543 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.943609 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-config-data\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.947036 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.947632 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-public-tls-certs\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:21 crc kubenswrapper[4862]: I1128 11:46:21.953901 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtlkq\" (UniqueName: \"kubernetes.io/projected/4a26f616-4cf2-441b-b61f-606edcb36a55-kube-api-access-dtlkq\") pod \"nova-api-0\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " pod="openstack/nova-api-0" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.015528 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.038061 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.038227 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.561383 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:22 crc kubenswrapper[4862]: W1128 11:46:22.569898 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a26f616_4cf2_441b_b61f_606edcb36a55.slice/crio-78d8776ab1540f54cf3ad8b78e01af4258eb38a69039369df30cb51907d03022 WatchSource:0}: Error finding container 78d8776ab1540f54cf3ad8b78e01af4258eb38a69039369df30cb51907d03022: Status 404 returned error can't find the container with id 78d8776ab1540f54cf3ad8b78e01af4258eb38a69039369df30cb51907d03022 Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.617748 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a26f616-4cf2-441b-b61f-606edcb36a55","Type":"ContainerStarted","Data":"78d8776ab1540f54cf3ad8b78e01af4258eb38a69039369df30cb51907d03022"} Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.635046 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.785791 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-t9gw8"] Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.787268 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.790019 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.790372 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.799867 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t9gw8"] Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.869750 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8c408c5-bcf9-4a70-8954-016ec8618502" path="/var/lib/kubelet/pods/e8c408c5-bcf9-4a70-8954-016ec8618502/volumes" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.981485 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-scripts\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.981559 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-config-data\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.981734 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:22 crc kubenswrapper[4862]: I1128 11:46:22.981772 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cv7v\" (UniqueName: \"kubernetes.io/projected/085bce11-bc08-44cf-a4db-d90046c891a9-kube-api-access-8cv7v\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.083629 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-config-data\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.083764 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.083803 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cv7v\" (UniqueName: \"kubernetes.io/projected/085bce11-bc08-44cf-a4db-d90046c891a9-kube-api-access-8cv7v\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.084324 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-scripts\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.088766 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-config-data\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.095891 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wrwbg" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerName="registry-server" probeResult="failure" output=< Nov 28 11:46:23 crc kubenswrapper[4862]: timeout: failed to connect service ":50051" within 1s Nov 28 11:46:23 crc kubenswrapper[4862]: > Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.096027 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-scripts\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.097534 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.105550 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cv7v\" (UniqueName: \"kubernetes.io/projected/085bce11-bc08-44cf-a4db-d90046c891a9-kube-api-access-8cv7v\") pod \"nova-cell1-cell-mapping-t9gw8\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.124882 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.581459 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t9gw8"] Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.632418 4862 generic.go:334] "Generic (PLEG): container finished" podID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerID="b6daeab2cb9e0eab83885f7c09a724e82f74ac5e24d33148a1716002cfa0c5fb" exitCode=0 Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.632500 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerDied","Data":"b6daeab2cb9e0eab83885f7c09a724e82f74ac5e24d33148a1716002cfa0c5fb"} Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.634157 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t9gw8" event={"ID":"085bce11-bc08-44cf-a4db-d90046c891a9","Type":"ContainerStarted","Data":"10ad894abc23a2708388e08382a3487cd2f6b3340774b4e74c61761d4b55f48a"} Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.636132 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a26f616-4cf2-441b-b61f-606edcb36a55","Type":"ContainerStarted","Data":"bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8"} Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.636180 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a26f616-4cf2-441b-b61f-606edcb36a55","Type":"ContainerStarted","Data":"2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6"} Nov 28 11:46:23 crc kubenswrapper[4862]: I1128 11:46:23.662071 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.662049313 podStartE2EDuration="2.662049313s" podCreationTimestamp="2025-11-28 11:46:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:46:23.653784806 +0000 UTC m=+1461.186298757" watchObservedRunningTime="2025-11-28 11:46:23.662049313 +0000 UTC m=+1461.194563224" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.038222 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.215291 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-ceilometer-tls-certs\") pod \"f90c0e88-8de7-41c0-a838-89df16f1024f\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.215382 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-log-httpd\") pod \"f90c0e88-8de7-41c0-a838-89df16f1024f\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.215417 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-scripts\") pod \"f90c0e88-8de7-41c0-a838-89df16f1024f\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.215496 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgvdj\" (UniqueName: \"kubernetes.io/projected/f90c0e88-8de7-41c0-a838-89df16f1024f-kube-api-access-qgvdj\") pod \"f90c0e88-8de7-41c0-a838-89df16f1024f\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.216360 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-run-httpd\") pod \"f90c0e88-8de7-41c0-a838-89df16f1024f\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.216419 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-config-data\") pod \"f90c0e88-8de7-41c0-a838-89df16f1024f\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.216450 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-combined-ca-bundle\") pod \"f90c0e88-8de7-41c0-a838-89df16f1024f\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.216538 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-sg-core-conf-yaml\") pod \"f90c0e88-8de7-41c0-a838-89df16f1024f\" (UID: \"f90c0e88-8de7-41c0-a838-89df16f1024f\") " Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.216683 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f90c0e88-8de7-41c0-a838-89df16f1024f" (UID: "f90c0e88-8de7-41c0-a838-89df16f1024f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.217036 4862 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.218038 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f90c0e88-8de7-41c0-a838-89df16f1024f" (UID: "f90c0e88-8de7-41c0-a838-89df16f1024f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.220311 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-scripts" (OuterVolumeSpecName: "scripts") pod "f90c0e88-8de7-41c0-a838-89df16f1024f" (UID: "f90c0e88-8de7-41c0-a838-89df16f1024f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.222994 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f90c0e88-8de7-41c0-a838-89df16f1024f-kube-api-access-qgvdj" (OuterVolumeSpecName: "kube-api-access-qgvdj") pod "f90c0e88-8de7-41c0-a838-89df16f1024f" (UID: "f90c0e88-8de7-41c0-a838-89df16f1024f"). InnerVolumeSpecName "kube-api-access-qgvdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.262426 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f90c0e88-8de7-41c0-a838-89df16f1024f" (UID: "f90c0e88-8de7-41c0-a838-89df16f1024f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.314595 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f90c0e88-8de7-41c0-a838-89df16f1024f" (UID: "f90c0e88-8de7-41c0-a838-89df16f1024f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.318893 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgvdj\" (UniqueName: \"kubernetes.io/projected/f90c0e88-8de7-41c0-a838-89df16f1024f-kube-api-access-qgvdj\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.318936 4862 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.318950 4862 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.318962 4862 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f90c0e88-8de7-41c0-a838-89df16f1024f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.318973 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.319789 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f90c0e88-8de7-41c0-a838-89df16f1024f" (UID: "f90c0e88-8de7-41c0-a838-89df16f1024f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.340255 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-config-data" (OuterVolumeSpecName: "config-data") pod "f90c0e88-8de7-41c0-a838-89df16f1024f" (UID: "f90c0e88-8de7-41c0-a838-89df16f1024f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.420847 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.421239 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90c0e88-8de7-41c0-a838-89df16f1024f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.655852 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f90c0e88-8de7-41c0-a838-89df16f1024f","Type":"ContainerDied","Data":"05aa8e1938d95ee4b9c720281bcbe39dea898b3038e6028e1143d21b764dfcae"} Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.655900 4862 scope.go:117] "RemoveContainer" containerID="4e3b49e4b2deac8f52253fb5568368b7416015ed50d5db8ec5c75aab2135ac13" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.656029 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.672009 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t9gw8" event={"ID":"085bce11-bc08-44cf-a4db-d90046c891a9","Type":"ContainerStarted","Data":"e3abc0084149b3ca5bb72b8e983fd6de4df9cf6de8997495203d2cf08dd8bfa8"} Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.686649 4862 scope.go:117] "RemoveContainer" containerID="a09438ad6c8003f94013130e6fbae489459ae08e8d3c237dd5a9dfd08caebdfd" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.694139 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-t9gw8" podStartSLOduration=2.694121186 podStartE2EDuration="2.694121186s" podCreationTimestamp="2025-11-28 11:46:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:46:24.690256177 +0000 UTC m=+1462.222770108" watchObservedRunningTime="2025-11-28 11:46:24.694121186 +0000 UTC m=+1462.226635117" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.722476 4862 scope.go:117] "RemoveContainer" containerID="b6daeab2cb9e0eab83885f7c09a724e82f74ac5e24d33148a1716002cfa0c5fb" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.739942 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.752859 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.761399 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:46:24 crc kubenswrapper[4862]: E1128 11:46:24.763131 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="proxy-httpd" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.763177 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="proxy-httpd" Nov 28 11:46:24 crc kubenswrapper[4862]: E1128 11:46:24.763245 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="ceilometer-central-agent" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.763255 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="ceilometer-central-agent" Nov 28 11:46:24 crc kubenswrapper[4862]: E1128 11:46:24.763287 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="ceilometer-notification-agent" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.763296 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="ceilometer-notification-agent" Nov 28 11:46:24 crc kubenswrapper[4862]: E1128 11:46:24.763320 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="sg-core" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.763336 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="sg-core" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.764303 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="ceilometer-notification-agent" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.764348 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="sg-core" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.764378 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="ceilometer-central-agent" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.764399 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" containerName="proxy-httpd" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.774338 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.779832 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.780067 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.780437 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.781059 4862 scope.go:117] "RemoveContainer" containerID="700a6ae64c6f634ec8de8f0e56d2e9e04f29aaf0eed950337541b42bcae36a57" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.796362 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.848541 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f90c0e88-8de7-41c0-a838-89df16f1024f" path="/var/lib/kubelet/pods/f90c0e88-8de7-41c0-a838-89df16f1024f/volumes" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.933331 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-log-httpd\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.933405 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-scripts\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.933471 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.933515 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.933540 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-run-httpd\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.933616 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-config-data\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.933650 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6gsx\" (UniqueName: \"kubernetes.io/projected/6b1b592a-d848-4857-894c-8f05a736dc6d-kube-api-access-x6gsx\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:24 crc kubenswrapper[4862]: I1128 11:46:24.933711 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.035544 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-log-httpd\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.035913 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-scripts\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.035946 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.035973 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.036006 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-run-httpd\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.036069 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-config-data\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.036116 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6gsx\" (UniqueName: \"kubernetes.io/projected/6b1b592a-d848-4857-894c-8f05a736dc6d-kube-api-access-x6gsx\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.036165 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.036577 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-log-httpd\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.038760 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-run-httpd\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.038770 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.044768 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.045050 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-config-data\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.045383 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.059284 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.078128 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6gsx\" (UniqueName: \"kubernetes.io/projected/6b1b592a-d848-4857-894c-8f05a736dc6d-kube-api-access-x6gsx\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.080957 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-scripts\") pod \"ceilometer-0\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.093802 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.134575 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75df6cf455-lxwbn"] Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.134815 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" podUID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" containerName="dnsmasq-dns" containerID="cri-o://3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c" gracePeriod=10 Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.635774 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:46:25 crc kubenswrapper[4862]: W1128 11:46:25.643251 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b1b592a_d848_4857_894c_8f05a736dc6d.slice/crio-edcb718a300c14df2b671048e5fbc7738c2ac9e67def205db08d60b1a0934fed WatchSource:0}: Error finding container edcb718a300c14df2b671048e5fbc7738c2ac9e67def205db08d60b1a0934fed: Status 404 returned error can't find the container with id edcb718a300c14df2b671048e5fbc7738c2ac9e67def205db08d60b1a0934fed Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.654841 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.683211 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerStarted","Data":"edcb718a300c14df2b671048e5fbc7738c2ac9e67def205db08d60b1a0934fed"} Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.690616 4862 generic.go:334] "Generic (PLEG): container finished" podID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" containerID="3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c" exitCode=0 Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.691251 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" event={"ID":"c4edcc93-fcf3-46a4-afe4-0df67274a67f","Type":"ContainerDied","Data":"3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c"} Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.691315 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" event={"ID":"c4edcc93-fcf3-46a4-afe4-0df67274a67f","Type":"ContainerDied","Data":"7a5eaf6915f3a5d400c748cfc7a14b5af21b248e596daf1ad5a3043c01903728"} Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.691327 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.691338 4862 scope.go:117] "RemoveContainer" containerID="3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.724865 4862 scope.go:117] "RemoveContainer" containerID="a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.753763 4862 scope.go:117] "RemoveContainer" containerID="3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c" Nov 28 11:46:25 crc kubenswrapper[4862]: E1128 11:46:25.754498 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c\": container with ID starting with 3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c not found: ID does not exist" containerID="3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.754544 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c"} err="failed to get container status \"3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c\": rpc error: code = NotFound desc = could not find container \"3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c\": container with ID starting with 3ac01a18bc0fdbe0b7b753ccb7a3c64a329b0df9da09bc7f6d9498e42ab1cd7c not found: ID does not exist" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.754570 4862 scope.go:117] "RemoveContainer" containerID="a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076" Nov 28 11:46:25 crc kubenswrapper[4862]: E1128 11:46:25.756645 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076\": container with ID starting with a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076 not found: ID does not exist" containerID="a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.756686 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076"} err="failed to get container status \"a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076\": rpc error: code = NotFound desc = could not find container \"a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076\": container with ID starting with a50a4fe8e99117da01b042c7f6ff722489cdd8516eb44a571fc11450bc154076 not found: ID does not exist" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.852658 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-nb\") pod \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.852950 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-sb\") pod \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.853002 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-svc\") pod \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.853043 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-config\") pod \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.853071 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-swift-storage-0\") pod \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.853126 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnlch\" (UniqueName: \"kubernetes.io/projected/c4edcc93-fcf3-46a4-afe4-0df67274a67f-kube-api-access-fnlch\") pod \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\" (UID: \"c4edcc93-fcf3-46a4-afe4-0df67274a67f\") " Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.859212 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4edcc93-fcf3-46a4-afe4-0df67274a67f-kube-api-access-fnlch" (OuterVolumeSpecName: "kube-api-access-fnlch") pod "c4edcc93-fcf3-46a4-afe4-0df67274a67f" (UID: "c4edcc93-fcf3-46a4-afe4-0df67274a67f"). InnerVolumeSpecName "kube-api-access-fnlch". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.904248 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-config" (OuterVolumeSpecName: "config") pod "c4edcc93-fcf3-46a4-afe4-0df67274a67f" (UID: "c4edcc93-fcf3-46a4-afe4-0df67274a67f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.915551 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c4edcc93-fcf3-46a4-afe4-0df67274a67f" (UID: "c4edcc93-fcf3-46a4-afe4-0df67274a67f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.921149 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c4edcc93-fcf3-46a4-afe4-0df67274a67f" (UID: "c4edcc93-fcf3-46a4-afe4-0df67274a67f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.930194 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4edcc93-fcf3-46a4-afe4-0df67274a67f" (UID: "c4edcc93-fcf3-46a4-afe4-0df67274a67f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.953662 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c4edcc93-fcf3-46a4-afe4-0df67274a67f" (UID: "c4edcc93-fcf3-46a4-afe4-0df67274a67f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.956834 4862 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.956881 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnlch\" (UniqueName: \"kubernetes.io/projected/c4edcc93-fcf3-46a4-afe4-0df67274a67f-kube-api-access-fnlch\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.956900 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.956912 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.956925 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:25 crc kubenswrapper[4862]: I1128 11:46:25.956938 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edcc93-fcf3-46a4-afe4-0df67274a67f-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:26 crc kubenswrapper[4862]: I1128 11:46:26.030029 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75df6cf455-lxwbn"] Nov 28 11:46:26 crc kubenswrapper[4862]: I1128 11:46:26.056342 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75df6cf455-lxwbn"] Nov 28 11:46:26 crc kubenswrapper[4862]: I1128 11:46:26.704949 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerStarted","Data":"11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945"} Nov 28 11:46:26 crc kubenswrapper[4862]: I1128 11:46:26.854861 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" path="/var/lib/kubelet/pods/c4edcc93-fcf3-46a4-afe4-0df67274a67f/volumes" Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.723581 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerStarted","Data":"f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe"} Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.871076 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s57qh"] Nov 28 11:46:27 crc kubenswrapper[4862]: E1128 11:46:27.871554 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" containerName="dnsmasq-dns" Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.871582 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" containerName="dnsmasq-dns" Nov 28 11:46:27 crc kubenswrapper[4862]: E1128 11:46:27.871619 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" containerName="init" Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.871629 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" containerName="init" Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.871896 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" containerName="dnsmasq-dns" Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.876561 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.896076 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s57qh"] Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.902718 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-catalog-content\") pod \"certified-operators-s57qh\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.902826 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzgxs\" (UniqueName: \"kubernetes.io/projected/b2edf8de-176a-44b4-96bf-ec3d59133569-kube-api-access-kzgxs\") pod \"certified-operators-s57qh\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:27 crc kubenswrapper[4862]: I1128 11:46:27.902867 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-utilities\") pod \"certified-operators-s57qh\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.010525 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzgxs\" (UniqueName: \"kubernetes.io/projected/b2edf8de-176a-44b4-96bf-ec3d59133569-kube-api-access-kzgxs\") pod \"certified-operators-s57qh\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.010582 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-utilities\") pod \"certified-operators-s57qh\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.010675 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-catalog-content\") pod \"certified-operators-s57qh\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.011391 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-utilities\") pod \"certified-operators-s57qh\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.011400 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-catalog-content\") pod \"certified-operators-s57qh\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.027759 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzgxs\" (UniqueName: \"kubernetes.io/projected/b2edf8de-176a-44b4-96bf-ec3d59133569-kube-api-access-kzgxs\") pod \"certified-operators-s57qh\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.210008 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.680935 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s57qh"] Nov 28 11:46:28 crc kubenswrapper[4862]: W1128 11:46:28.684005 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2edf8de_176a_44b4_96bf_ec3d59133569.slice/crio-b31d5a0b736b30a57a50dc8c6f37a2a26a5d141dd1ad43a3babd1edaacfc0083 WatchSource:0}: Error finding container b31d5a0b736b30a57a50dc8c6f37a2a26a5d141dd1ad43a3babd1edaacfc0083: Status 404 returned error can't find the container with id b31d5a0b736b30a57a50dc8c6f37a2a26a5d141dd1ad43a3babd1edaacfc0083 Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.742169 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s57qh" event={"ID":"b2edf8de-176a-44b4-96bf-ec3d59133569","Type":"ContainerStarted","Data":"b31d5a0b736b30a57a50dc8c6f37a2a26a5d141dd1ad43a3babd1edaacfc0083"} Nov 28 11:46:28 crc kubenswrapper[4862]: I1128 11:46:28.746179 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerStarted","Data":"eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee"} Nov 28 11:46:29 crc kubenswrapper[4862]: I1128 11:46:29.759397 4862 generic.go:334] "Generic (PLEG): container finished" podID="085bce11-bc08-44cf-a4db-d90046c891a9" containerID="e3abc0084149b3ca5bb72b8e983fd6de4df9cf6de8997495203d2cf08dd8bfa8" exitCode=0 Nov 28 11:46:29 crc kubenswrapper[4862]: I1128 11:46:29.759632 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t9gw8" event={"ID":"085bce11-bc08-44cf-a4db-d90046c891a9","Type":"ContainerDied","Data":"e3abc0084149b3ca5bb72b8e983fd6de4df9cf6de8997495203d2cf08dd8bfa8"} Nov 28 11:46:29 crc kubenswrapper[4862]: I1128 11:46:29.762112 4862 generic.go:334] "Generic (PLEG): container finished" podID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerID="643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3" exitCode=0 Nov 28 11:46:29 crc kubenswrapper[4862]: I1128 11:46:29.762154 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s57qh" event={"ID":"b2edf8de-176a-44b4-96bf-ec3d59133569","Type":"ContainerDied","Data":"643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3"} Nov 28 11:46:30 crc kubenswrapper[4862]: I1128 11:46:30.602188 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-75df6cf455-lxwbn" podUID="c4edcc93-fcf3-46a4-afe4-0df67274a67f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.184:5353: i/o timeout" Nov 28 11:46:30 crc kubenswrapper[4862]: I1128 11:46:30.775700 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s57qh" event={"ID":"b2edf8de-176a-44b4-96bf-ec3d59133569","Type":"ContainerStarted","Data":"65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5"} Nov 28 11:46:30 crc kubenswrapper[4862]: I1128 11:46:30.781605 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerStarted","Data":"35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde"} Nov 28 11:46:30 crc kubenswrapper[4862]: I1128 11:46:30.838942 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.938946953 podStartE2EDuration="6.838918052s" podCreationTimestamp="2025-11-28 11:46:24 +0000 UTC" firstStartedPulling="2025-11-28 11:46:25.644964872 +0000 UTC m=+1463.177478793" lastFinishedPulling="2025-11-28 11:46:29.544935961 +0000 UTC m=+1467.077449892" observedRunningTime="2025-11-28 11:46:30.833198978 +0000 UTC m=+1468.365712899" watchObservedRunningTime="2025-11-28 11:46:30.838918052 +0000 UTC m=+1468.371431973" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.182267 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.288288 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-scripts\") pod \"085bce11-bc08-44cf-a4db-d90046c891a9\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.288383 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cv7v\" (UniqueName: \"kubernetes.io/projected/085bce11-bc08-44cf-a4db-d90046c891a9-kube-api-access-8cv7v\") pod \"085bce11-bc08-44cf-a4db-d90046c891a9\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.288488 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-combined-ca-bundle\") pod \"085bce11-bc08-44cf-a4db-d90046c891a9\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.288533 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-config-data\") pod \"085bce11-bc08-44cf-a4db-d90046c891a9\" (UID: \"085bce11-bc08-44cf-a4db-d90046c891a9\") " Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.294909 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085bce11-bc08-44cf-a4db-d90046c891a9-kube-api-access-8cv7v" (OuterVolumeSpecName: "kube-api-access-8cv7v") pod "085bce11-bc08-44cf-a4db-d90046c891a9" (UID: "085bce11-bc08-44cf-a4db-d90046c891a9"). InnerVolumeSpecName "kube-api-access-8cv7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.295656 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-scripts" (OuterVolumeSpecName: "scripts") pod "085bce11-bc08-44cf-a4db-d90046c891a9" (UID: "085bce11-bc08-44cf-a4db-d90046c891a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.334838 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "085bce11-bc08-44cf-a4db-d90046c891a9" (UID: "085bce11-bc08-44cf-a4db-d90046c891a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.351080 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-config-data" (OuterVolumeSpecName: "config-data") pod "085bce11-bc08-44cf-a4db-d90046c891a9" (UID: "085bce11-bc08-44cf-a4db-d90046c891a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.389763 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.389790 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cv7v\" (UniqueName: \"kubernetes.io/projected/085bce11-bc08-44cf-a4db-d90046c891a9-kube-api-access-8cv7v\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.389801 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.389811 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085bce11-bc08-44cf-a4db-d90046c891a9-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.806877 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t9gw8" event={"ID":"085bce11-bc08-44cf-a4db-d90046c891a9","Type":"ContainerDied","Data":"10ad894abc23a2708388e08382a3487cd2f6b3340774b4e74c61761d4b55f48a"} Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.806941 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t9gw8" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.806949 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10ad894abc23a2708388e08382a3487cd2f6b3340774b4e74c61761d4b55f48a" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.810079 4862 generic.go:334] "Generic (PLEG): container finished" podID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerID="65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5" exitCode=0 Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.813235 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s57qh" event={"ID":"b2edf8de-176a-44b4-96bf-ec3d59133569","Type":"ContainerDied","Data":"65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5"} Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.813338 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.983623 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.984485 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerName="nova-api-api" containerID="cri-o://bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8" gracePeriod=30 Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.984304 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerName="nova-api-log" containerID="cri-o://2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6" gracePeriod=30 Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.998753 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:46:31 crc kubenswrapper[4862]: I1128 11:46:31.999002 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b1edabad-9cf9-4e4f-935a-01493ebce5ec" containerName="nova-scheduler-scheduler" containerID="cri-o://921db04cc1e5a5fe187dff1e200d67d69b50f9c36daae90f7a96a302f3ad65cc" gracePeriod=30 Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.050521 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.050778 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-log" containerID="cri-o://4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0" gracePeriod=30 Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.051265 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-metadata" containerID="cri-o://a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49" gracePeriod=30 Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.094922 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.161635 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:32 crc kubenswrapper[4862]: E1128 11:46:32.507242 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="921db04cc1e5a5fe187dff1e200d67d69b50f9c36daae90f7a96a302f3ad65cc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 11:46:32 crc kubenswrapper[4862]: E1128 11:46:32.520431 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="921db04cc1e5a5fe187dff1e200d67d69b50f9c36daae90f7a96a302f3ad65cc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 11:46:32 crc kubenswrapper[4862]: E1128 11:46:32.524208 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="921db04cc1e5a5fe187dff1e200d67d69b50f9c36daae90f7a96a302f3ad65cc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 11:46:32 crc kubenswrapper[4862]: E1128 11:46:32.524239 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b1edabad-9cf9-4e4f-935a-01493ebce5ec" containerName="nova-scheduler-scheduler" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.721467 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.738372 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtlkq\" (UniqueName: \"kubernetes.io/projected/4a26f616-4cf2-441b-b61f-606edcb36a55-kube-api-access-dtlkq\") pod \"4a26f616-4cf2-441b-b61f-606edcb36a55\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.738573 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-combined-ca-bundle\") pod \"4a26f616-4cf2-441b-b61f-606edcb36a55\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.738611 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-config-data\") pod \"4a26f616-4cf2-441b-b61f-606edcb36a55\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.738648 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-public-tls-certs\") pod \"4a26f616-4cf2-441b-b61f-606edcb36a55\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.738704 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a26f616-4cf2-441b-b61f-606edcb36a55-logs\") pod \"4a26f616-4cf2-441b-b61f-606edcb36a55\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.738750 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-internal-tls-certs\") pod \"4a26f616-4cf2-441b-b61f-606edcb36a55\" (UID: \"4a26f616-4cf2-441b-b61f-606edcb36a55\") " Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.740485 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a26f616-4cf2-441b-b61f-606edcb36a55-logs" (OuterVolumeSpecName: "logs") pod "4a26f616-4cf2-441b-b61f-606edcb36a55" (UID: "4a26f616-4cf2-441b-b61f-606edcb36a55"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.744337 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a26f616-4cf2-441b-b61f-606edcb36a55-kube-api-access-dtlkq" (OuterVolumeSpecName: "kube-api-access-dtlkq") pod "4a26f616-4cf2-441b-b61f-606edcb36a55" (UID: "4a26f616-4cf2-441b-b61f-606edcb36a55"). InnerVolumeSpecName "kube-api-access-dtlkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.793359 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-config-data" (OuterVolumeSpecName: "config-data") pod "4a26f616-4cf2-441b-b61f-606edcb36a55" (UID: "4a26f616-4cf2-441b-b61f-606edcb36a55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.796679 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a26f616-4cf2-441b-b61f-606edcb36a55" (UID: "4a26f616-4cf2-441b-b61f-606edcb36a55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.807043 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4a26f616-4cf2-441b-b61f-606edcb36a55" (UID: "4a26f616-4cf2-441b-b61f-606edcb36a55"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.815376 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4a26f616-4cf2-441b-b61f-606edcb36a55" (UID: "4a26f616-4cf2-441b-b61f-606edcb36a55"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.823148 4862 generic.go:334] "Generic (PLEG): container finished" podID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerID="bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8" exitCode=0 Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.823193 4862 generic.go:334] "Generic (PLEG): container finished" podID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerID="2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6" exitCode=143 Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.823237 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a26f616-4cf2-441b-b61f-606edcb36a55","Type":"ContainerDied","Data":"bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8"} Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.823266 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a26f616-4cf2-441b-b61f-606edcb36a55","Type":"ContainerDied","Data":"2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6"} Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.823276 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a26f616-4cf2-441b-b61f-606edcb36a55","Type":"ContainerDied","Data":"78d8776ab1540f54cf3ad8b78e01af4258eb38a69039369df30cb51907d03022"} Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.823290 4862 scope.go:117] "RemoveContainer" containerID="bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.823463 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.831037 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s57qh" event={"ID":"b2edf8de-176a-44b4-96bf-ec3d59133569","Type":"ContainerStarted","Data":"f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e"} Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.834154 4862 generic.go:334] "Generic (PLEG): container finished" podID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerID="4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0" exitCode=143 Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.834782 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22b82b62-3276-4527-90b3-d2abbbbb92ab","Type":"ContainerDied","Data":"4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0"} Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.841531 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a26f616-4cf2-441b-b61f-606edcb36a55-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.841560 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.841571 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtlkq\" (UniqueName: \"kubernetes.io/projected/4a26f616-4cf2-441b-b61f-606edcb36a55-kube-api-access-dtlkq\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.841580 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.841590 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.841600 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a26f616-4cf2-441b-b61f-606edcb36a55-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.853070 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s57qh" podStartSLOduration=3.337005388 podStartE2EDuration="5.853029614s" podCreationTimestamp="2025-11-28 11:46:27 +0000 UTC" firstStartedPulling="2025-11-28 11:46:29.766211715 +0000 UTC m=+1467.298725636" lastFinishedPulling="2025-11-28 11:46:32.282235941 +0000 UTC m=+1469.814749862" observedRunningTime="2025-11-28 11:46:32.851761027 +0000 UTC m=+1470.384274948" watchObservedRunningTime="2025-11-28 11:46:32.853029614 +0000 UTC m=+1470.385543535" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.870356 4862 scope.go:117] "RemoveContainer" containerID="2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.899958 4862 scope.go:117] "RemoveContainer" containerID="bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8" Nov 28 11:46:32 crc kubenswrapper[4862]: E1128 11:46:32.901583 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8\": container with ID starting with bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8 not found: ID does not exist" containerID="bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.901688 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8"} err="failed to get container status \"bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8\": rpc error: code = NotFound desc = could not find container \"bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8\": container with ID starting with bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8 not found: ID does not exist" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.901788 4862 scope.go:117] "RemoveContainer" containerID="2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6" Nov 28 11:46:32 crc kubenswrapper[4862]: E1128 11:46:32.902201 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6\": container with ID starting with 2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6 not found: ID does not exist" containerID="2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.902259 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6"} err="failed to get container status \"2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6\": rpc error: code = NotFound desc = could not find container \"2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6\": container with ID starting with 2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6 not found: ID does not exist" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.902295 4862 scope.go:117] "RemoveContainer" containerID="bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.902405 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.902845 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8"} err="failed to get container status \"bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8\": rpc error: code = NotFound desc = could not find container \"bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8\": container with ID starting with bcf42cb90ad27d96bc9819f4e0d408e47ac878be1f6c411273c88c0fd0550ee8 not found: ID does not exist" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.902875 4862 scope.go:117] "RemoveContainer" containerID="2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.903114 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6"} err="failed to get container status \"2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6\": rpc error: code = NotFound desc = could not find container \"2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6\": container with ID starting with 2b5f78d5041505a54354eebaab7c6eb653428b39f46b4530cb417dcc923d69d6 not found: ID does not exist" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.916959 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.925000 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:32 crc kubenswrapper[4862]: E1128 11:46:32.925403 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerName="nova-api-log" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.925420 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerName="nova-api-log" Nov 28 11:46:32 crc kubenswrapper[4862]: E1128 11:46:32.925434 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerName="nova-api-api" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.925441 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerName="nova-api-api" Nov 28 11:46:32 crc kubenswrapper[4862]: E1128 11:46:32.925453 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085bce11-bc08-44cf-a4db-d90046c891a9" containerName="nova-manage" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.925459 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="085bce11-bc08-44cf-a4db-d90046c891a9" containerName="nova-manage" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.925662 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="085bce11-bc08-44cf-a4db-d90046c891a9" containerName="nova-manage" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.925681 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerName="nova-api-log" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.925690 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a26f616-4cf2-441b-b61f-606edcb36a55" containerName="nova-api-api" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.926652 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.929033 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.929598 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.929783 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.942799 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.943079 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77de7245-a8de-4912-b90f-669c57d8d563-logs\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.943131 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-public-tls-certs\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.943254 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-internal-tls-certs\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.943328 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t2jd\" (UniqueName: \"kubernetes.io/projected/77de7245-a8de-4912-b90f-669c57d8d563-kube-api-access-9t2jd\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.943384 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:32 crc kubenswrapper[4862]: I1128 11:46:32.943425 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-config-data\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.045413 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-config-data\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.046429 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77de7245-a8de-4912-b90f-669c57d8d563-logs\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.046954 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-public-tls-certs\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.047277 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-internal-tls-certs\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.047423 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t2jd\" (UniqueName: \"kubernetes.io/projected/77de7245-a8de-4912-b90f-669c57d8d563-kube-api-access-9t2jd\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.047556 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.047563 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77de7245-a8de-4912-b90f-669c57d8d563-logs\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.053895 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-config-data\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.055293 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-internal-tls-certs\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.056442 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.056535 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-public-tls-certs\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.066708 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t2jd\" (UniqueName: \"kubernetes.io/projected/77de7245-a8de-4912-b90f-669c57d8d563-kube-api-access-9t2jd\") pod \"nova-api-0\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.246388 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.741325 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:46:33 crc kubenswrapper[4862]: I1128 11:46:33.849813 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77de7245-a8de-4912-b90f-669c57d8d563","Type":"ContainerStarted","Data":"2fac0b3dcfcd10093df99f352ea83a49a5118ccdde56ae84f862b2f3813021dd"} Nov 28 11:46:34 crc kubenswrapper[4862]: I1128 11:46:34.446792 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wrwbg"] Nov 28 11:46:34 crc kubenswrapper[4862]: I1128 11:46:34.447405 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wrwbg" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerName="registry-server" containerID="cri-o://bafdf6e87c9e4c8287da1ee020aafbb27a9b6879dd62516aa2058138e4476799" gracePeriod=2 Nov 28 11:46:34 crc kubenswrapper[4862]: I1128 11:46:34.868589 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a26f616-4cf2-441b-b61f-606edcb36a55" path="/var/lib/kubelet/pods/4a26f616-4cf2-441b-b61f-606edcb36a55/volumes" Nov 28 11:46:34 crc kubenswrapper[4862]: I1128 11:46:34.869605 4862 generic.go:334] "Generic (PLEG): container finished" podID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerID="bafdf6e87c9e4c8287da1ee020aafbb27a9b6879dd62516aa2058138e4476799" exitCode=0 Nov 28 11:46:34 crc kubenswrapper[4862]: I1128 11:46:34.869697 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77de7245-a8de-4912-b90f-669c57d8d563","Type":"ContainerStarted","Data":"69388439d162f0fc46186cc6f4708492782333d0c2f8c07c9238c3381aaf99ad"} Nov 28 11:46:34 crc kubenswrapper[4862]: I1128 11:46:34.869750 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77de7245-a8de-4912-b90f-669c57d8d563","Type":"ContainerStarted","Data":"adb196b1aa7dfec6e01189bb208acacebcbed93cfaed7fde33e15993b31a31f9"} Nov 28 11:46:34 crc kubenswrapper[4862]: I1128 11:46:34.869764 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrwbg" event={"ID":"4b3b2737-87e3-4017-965e-3a49a88c4526","Type":"ContainerDied","Data":"bafdf6e87c9e4c8287da1ee020aafbb27a9b6879dd62516aa2058138e4476799"} Nov 28 11:46:34 crc kubenswrapper[4862]: I1128 11:46:34.898864 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8988402239999997 podStartE2EDuration="2.898840224s" podCreationTimestamp="2025-11-28 11:46:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:46:34.889964239 +0000 UTC m=+1472.422478160" watchObservedRunningTime="2025-11-28 11:46:34.898840224 +0000 UTC m=+1472.431354185" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.004901 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.100448 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-catalog-content\") pod \"4b3b2737-87e3-4017-965e-3a49a88c4526\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.100785 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-utilities\") pod \"4b3b2737-87e3-4017-965e-3a49a88c4526\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.100863 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnczv\" (UniqueName: \"kubernetes.io/projected/4b3b2737-87e3-4017-965e-3a49a88c4526-kube-api-access-dnczv\") pod \"4b3b2737-87e3-4017-965e-3a49a88c4526\" (UID: \"4b3b2737-87e3-4017-965e-3a49a88c4526\") " Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.102809 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-utilities" (OuterVolumeSpecName: "utilities") pod "4b3b2737-87e3-4017-965e-3a49a88c4526" (UID: "4b3b2737-87e3-4017-965e-3a49a88c4526"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.106869 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b3b2737-87e3-4017-965e-3a49a88c4526-kube-api-access-dnczv" (OuterVolumeSpecName: "kube-api-access-dnczv") pod "4b3b2737-87e3-4017-965e-3a49a88c4526" (UID: "4b3b2737-87e3-4017-965e-3a49a88c4526"). InnerVolumeSpecName "kube-api-access-dnczv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.202754 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.202792 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnczv\" (UniqueName: \"kubernetes.io/projected/4b3b2737-87e3-4017-965e-3a49a88c4526-kube-api-access-dnczv\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.223638 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b3b2737-87e3-4017-965e-3a49a88c4526" (UID: "4b3b2737-87e3-4017-965e-3a49a88c4526"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.304929 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b3b2737-87e3-4017-965e-3a49a88c4526-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.680640 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.711308 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b82b62-3276-4527-90b3-d2abbbbb92ab-logs\") pod \"22b82b62-3276-4527-90b3-d2abbbbb92ab\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.711360 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-config-data\") pod \"22b82b62-3276-4527-90b3-d2abbbbb92ab\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.711413 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6d26\" (UniqueName: \"kubernetes.io/projected/22b82b62-3276-4527-90b3-d2abbbbb92ab-kube-api-access-c6d26\") pod \"22b82b62-3276-4527-90b3-d2abbbbb92ab\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.711514 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-nova-metadata-tls-certs\") pod \"22b82b62-3276-4527-90b3-d2abbbbb92ab\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.711567 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-combined-ca-bundle\") pod \"22b82b62-3276-4527-90b3-d2abbbbb92ab\" (UID: \"22b82b62-3276-4527-90b3-d2abbbbb92ab\") " Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.711889 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22b82b62-3276-4527-90b3-d2abbbbb92ab-logs" (OuterVolumeSpecName: "logs") pod "22b82b62-3276-4527-90b3-d2abbbbb92ab" (UID: "22b82b62-3276-4527-90b3-d2abbbbb92ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.712244 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b82b62-3276-4527-90b3-d2abbbbb92ab-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.715629 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22b82b62-3276-4527-90b3-d2abbbbb92ab-kube-api-access-c6d26" (OuterVolumeSpecName: "kube-api-access-c6d26") pod "22b82b62-3276-4527-90b3-d2abbbbb92ab" (UID: "22b82b62-3276-4527-90b3-d2abbbbb92ab"). InnerVolumeSpecName "kube-api-access-c6d26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.750552 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-config-data" (OuterVolumeSpecName: "config-data") pod "22b82b62-3276-4527-90b3-d2abbbbb92ab" (UID: "22b82b62-3276-4527-90b3-d2abbbbb92ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.784398 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22b82b62-3276-4527-90b3-d2abbbbb92ab" (UID: "22b82b62-3276-4527-90b3-d2abbbbb92ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.809958 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "22b82b62-3276-4527-90b3-d2abbbbb92ab" (UID: "22b82b62-3276-4527-90b3-d2abbbbb92ab"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.814526 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6d26\" (UniqueName: \"kubernetes.io/projected/22b82b62-3276-4527-90b3-d2abbbbb92ab-kube-api-access-c6d26\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.814560 4862 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.814576 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.814589 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b82b62-3276-4527-90b3-d2abbbbb92ab-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.888130 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrwbg" event={"ID":"4b3b2737-87e3-4017-965e-3a49a88c4526","Type":"ContainerDied","Data":"51af13f1d7e96a7dd7217b53a3143b46046d278bb49ce435bb883d7060d5d5cb"} Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.888185 4862 scope.go:117] "RemoveContainer" containerID="bafdf6e87c9e4c8287da1ee020aafbb27a9b6879dd62516aa2058138e4476799" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.888208 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wrwbg" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.894737 4862 generic.go:334] "Generic (PLEG): container finished" podID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerID="a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49" exitCode=0 Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.894947 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.895176 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22b82b62-3276-4527-90b3-d2abbbbb92ab","Type":"ContainerDied","Data":"a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49"} Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.895210 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22b82b62-3276-4527-90b3-d2abbbbb92ab","Type":"ContainerDied","Data":"99da844c8d25b38a0dfe38d51d4c8b8c1f73ebc1724f101db4a0cbb63126998f"} Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.910128 4862 scope.go:117] "RemoveContainer" containerID="f1560a3f2403b0cebf9942cbcd616447fa7c77edc5218f4969d1d49654b4431d" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.950073 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.950206 4862 scope.go:117] "RemoveContainer" containerID="597bcb30c2ec13a42a442bccc8d17ef41baa8183434146aff53ffd1727940883" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.965973 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.979071 4862 scope.go:117] "RemoveContainer" containerID="a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49" Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.982715 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wrwbg"] Nov 28 11:46:35 crc kubenswrapper[4862]: I1128 11:46:35.996715 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wrwbg"] Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.009650 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:46:36 crc kubenswrapper[4862]: E1128 11:46:36.010161 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-metadata" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.010186 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-metadata" Nov 28 11:46:36 crc kubenswrapper[4862]: E1128 11:46:36.010209 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerName="extract-content" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.010218 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerName="extract-content" Nov 28 11:46:36 crc kubenswrapper[4862]: E1128 11:46:36.010240 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerName="registry-server" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.010248 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerName="registry-server" Nov 28 11:46:36 crc kubenswrapper[4862]: E1128 11:46:36.010264 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerName="extract-utilities" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.010273 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerName="extract-utilities" Nov 28 11:46:36 crc kubenswrapper[4862]: E1128 11:46:36.010304 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-log" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.010312 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-log" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.010536 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" containerName="registry-server" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.010566 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-log" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.010586 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-metadata" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.011905 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.014010 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.014058 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.025177 4862 scope.go:117] "RemoveContainer" containerID="4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.050074 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.057308 4862 scope.go:117] "RemoveContainer" containerID="a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49" Nov 28 11:46:36 crc kubenswrapper[4862]: E1128 11:46:36.058115 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49\": container with ID starting with a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49 not found: ID does not exist" containerID="a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.058170 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49"} err="failed to get container status \"a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49\": rpc error: code = NotFound desc = could not find container \"a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49\": container with ID starting with a9cf92c22ad60521d18251b32a6cbfc5cbf41e66d7c0accff542d209fae31c49 not found: ID does not exist" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.058202 4862 scope.go:117] "RemoveContainer" containerID="4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0" Nov 28 11:46:36 crc kubenswrapper[4862]: E1128 11:46:36.058535 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0\": container with ID starting with 4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0 not found: ID does not exist" containerID="4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.058559 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0"} err="failed to get container status \"4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0\": rpc error: code = NotFound desc = could not find container \"4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0\": container with ID starting with 4ef5b606e5282546057b64a232c12b4281e8747034c7d7ebd577282ff0e5a2f0 not found: ID does not exist" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.119642 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-config-data\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.119711 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ed7ff-0788-436d-b811-a4aafc7f9f5f-logs\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.119747 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.119802 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.119829 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b584j\" (UniqueName: \"kubernetes.io/projected/310ed7ff-0788-436d-b811-a4aafc7f9f5f-kube-api-access-b584j\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.221327 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-config-data\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.221369 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ed7ff-0788-436d-b811-a4aafc7f9f5f-logs\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.221405 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.221461 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.221487 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b584j\" (UniqueName: \"kubernetes.io/projected/310ed7ff-0788-436d-b811-a4aafc7f9f5f-kube-api-access-b584j\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.224607 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ed7ff-0788-436d-b811-a4aafc7f9f5f-logs\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.229339 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-config-data\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.230469 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.232431 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.243010 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b584j\" (UniqueName: \"kubernetes.io/projected/310ed7ff-0788-436d-b811-a4aafc7f9f5f-kube-api-access-b584j\") pod \"nova-metadata-0\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.340955 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.865792 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" path="/var/lib/kubelet/pods/22b82b62-3276-4527-90b3-d2abbbbb92ab/volumes" Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.867851 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b3b2737-87e3-4017-965e-3a49a88c4526" path="/var/lib/kubelet/pods/4b3b2737-87e3-4017-965e-3a49a88c4526/volumes" Nov 28 11:46:36 crc kubenswrapper[4862]: W1128 11:46:36.900987 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod310ed7ff_0788_436d_b811_a4aafc7f9f5f.slice/crio-b9548b98bbb6b54ca79db5174ace520be3e184d53112db741e8ed1cd25393c9a WatchSource:0}: Error finding container b9548b98bbb6b54ca79db5174ace520be3e184d53112db741e8ed1cd25393c9a: Status 404 returned error can't find the container with id b9548b98bbb6b54ca79db5174ace520be3e184d53112db741e8ed1cd25393c9a Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.901678 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.915731 4862 generic.go:334] "Generic (PLEG): container finished" podID="b1edabad-9cf9-4e4f-935a-01493ebce5ec" containerID="921db04cc1e5a5fe187dff1e200d67d69b50f9c36daae90f7a96a302f3ad65cc" exitCode=0 Nov 28 11:46:36 crc kubenswrapper[4862]: I1128 11:46:36.915781 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1edabad-9cf9-4e4f-935a-01493ebce5ec","Type":"ContainerDied","Data":"921db04cc1e5a5fe187dff1e200d67d69b50f9c36daae90f7a96a302f3ad65cc"} Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.336917 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.446149 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-config-data\") pod \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.446243 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-combined-ca-bundle\") pod \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.446352 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8fpg\" (UniqueName: \"kubernetes.io/projected/b1edabad-9cf9-4e4f-935a-01493ebce5ec-kube-api-access-h8fpg\") pod \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\" (UID: \"b1edabad-9cf9-4e4f-935a-01493ebce5ec\") " Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.449760 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1edabad-9cf9-4e4f-935a-01493ebce5ec-kube-api-access-h8fpg" (OuterVolumeSpecName: "kube-api-access-h8fpg") pod "b1edabad-9cf9-4e4f-935a-01493ebce5ec" (UID: "b1edabad-9cf9-4e4f-935a-01493ebce5ec"). InnerVolumeSpecName "kube-api-access-h8fpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.470648 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1edabad-9cf9-4e4f-935a-01493ebce5ec" (UID: "b1edabad-9cf9-4e4f-935a-01493ebce5ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.477280 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-config-data" (OuterVolumeSpecName: "config-data") pod "b1edabad-9cf9-4e4f-935a-01493ebce5ec" (UID: "b1edabad-9cf9-4e4f-935a-01493ebce5ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.549119 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.549159 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1edabad-9cf9-4e4f-935a-01493ebce5ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.549174 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8fpg\" (UniqueName: \"kubernetes.io/projected/b1edabad-9cf9-4e4f-935a-01493ebce5ec-kube-api-access-h8fpg\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.927266 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"310ed7ff-0788-436d-b811-a4aafc7f9f5f","Type":"ContainerStarted","Data":"15217408eaf0c5ecfdbf05adb57c81819bf35870f42f21b0ecf1bd8c4c5ff8a8"} Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.927624 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"310ed7ff-0788-436d-b811-a4aafc7f9f5f","Type":"ContainerStarted","Data":"a303ac97d092846a039471dffa93b8705481508ae03493b999176d9423f8a8c4"} Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.927633 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"310ed7ff-0788-436d-b811-a4aafc7f9f5f","Type":"ContainerStarted","Data":"b9548b98bbb6b54ca79db5174ace520be3e184d53112db741e8ed1cd25393c9a"} Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.930668 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1edabad-9cf9-4e4f-935a-01493ebce5ec","Type":"ContainerDied","Data":"7c5f846f1d287be1a630204bfc75fc1714b01a082519ce675de2d59fada36438"} Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.930744 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.930758 4862 scope.go:117] "RemoveContainer" containerID="921db04cc1e5a5fe187dff1e200d67d69b50f9c36daae90f7a96a302f3ad65cc" Nov 28 11:46:37 crc kubenswrapper[4862]: I1128 11:46:37.963498 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.963469536 podStartE2EDuration="2.963469536s" podCreationTimestamp="2025-11-28 11:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:46:37.947834044 +0000 UTC m=+1475.480347985" watchObservedRunningTime="2025-11-28 11:46:37.963469536 +0000 UTC m=+1475.495983497" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.016370 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.028049 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.040598 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:46:38 crc kubenswrapper[4862]: E1128 11:46:38.041039 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1edabad-9cf9-4e4f-935a-01493ebce5ec" containerName="nova-scheduler-scheduler" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.041058 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1edabad-9cf9-4e4f-935a-01493ebce5ec" containerName="nova-scheduler-scheduler" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.041327 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1edabad-9cf9-4e4f-935a-01493ebce5ec" containerName="nova-scheduler-scheduler" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.041960 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.048531 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.050713 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.063678 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-config-data\") pod \"nova-scheduler-0\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.063735 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.063757 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf7gr\" (UniqueName: \"kubernetes.io/projected/a824cc2c-2812-4d91-9e9f-9736a27fe794-kube-api-access-kf7gr\") pod \"nova-scheduler-0\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.165563 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-config-data\") pod \"nova-scheduler-0\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.165624 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.165644 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf7gr\" (UniqueName: \"kubernetes.io/projected/a824cc2c-2812-4d91-9e9f-9736a27fe794-kube-api-access-kf7gr\") pod \"nova-scheduler-0\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.170937 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.171983 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-config-data\") pod \"nova-scheduler-0\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.188392 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf7gr\" (UniqueName: \"kubernetes.io/projected/a824cc2c-2812-4d91-9e9f-9736a27fe794-kube-api-access-kf7gr\") pod \"nova-scheduler-0\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.210900 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.210955 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.281077 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.368673 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.848343 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1edabad-9cf9-4e4f-935a-01493ebce5ec" path="/var/lib/kubelet/pods/b1edabad-9cf9-4e4f-935a-01493ebce5ec/volumes" Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.928276 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:46:38 crc kubenswrapper[4862]: W1128 11:46:38.930276 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda824cc2c_2812_4d91_9e9f_9736a27fe794.slice/crio-d10eaea45792fe4cd29893a74175c64063f071faa3833b0635963d49903e4e16 WatchSource:0}: Error finding container d10eaea45792fe4cd29893a74175c64063f071faa3833b0635963d49903e4e16: Status 404 returned error can't find the container with id d10eaea45792fe4cd29893a74175c64063f071faa3833b0635963d49903e4e16 Nov 28 11:46:38 crc kubenswrapper[4862]: I1128 11:46:38.942835 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a824cc2c-2812-4d91-9e9f-9736a27fe794","Type":"ContainerStarted","Data":"d10eaea45792fe4cd29893a74175c64063f071faa3833b0635963d49903e4e16"} Nov 28 11:46:39 crc kubenswrapper[4862]: I1128 11:46:39.025575 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:39 crc kubenswrapper[4862]: I1128 11:46:39.454181 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s57qh"] Nov 28 11:46:39 crc kubenswrapper[4862]: I1128 11:46:39.963653 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a824cc2c-2812-4d91-9e9f-9736a27fe794","Type":"ContainerStarted","Data":"0682b22558ac57d011862a3e8a94d6b4334e16c0f890d30fb6ae86e9faaac01e"} Nov 28 11:46:40 crc kubenswrapper[4862]: I1128 11:46:40.015042 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.015020611 podStartE2EDuration="3.015020611s" podCreationTimestamp="2025-11-28 11:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:46:39.998059042 +0000 UTC m=+1477.530573003" watchObservedRunningTime="2025-11-28 11:46:40.015020611 +0000 UTC m=+1477.547534542" Nov 28 11:46:40 crc kubenswrapper[4862]: I1128 11:46:40.513341 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:40 crc kubenswrapper[4862]: I1128 11:46:40.513326 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="22b82b62-3276-4527-90b3-d2abbbbb92ab" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:40 crc kubenswrapper[4862]: I1128 11:46:40.976018 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s57qh" podUID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerName="registry-server" containerID="cri-o://f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e" gracePeriod=2 Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.341444 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.341856 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.488175 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.544617 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-utilities\") pod \"b2edf8de-176a-44b4-96bf-ec3d59133569\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.544829 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-catalog-content\") pod \"b2edf8de-176a-44b4-96bf-ec3d59133569\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.544967 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzgxs\" (UniqueName: \"kubernetes.io/projected/b2edf8de-176a-44b4-96bf-ec3d59133569-kube-api-access-kzgxs\") pod \"b2edf8de-176a-44b4-96bf-ec3d59133569\" (UID: \"b2edf8de-176a-44b4-96bf-ec3d59133569\") " Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.546009 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-utilities" (OuterVolumeSpecName: "utilities") pod "b2edf8de-176a-44b4-96bf-ec3d59133569" (UID: "b2edf8de-176a-44b4-96bf-ec3d59133569"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.559392 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2edf8de-176a-44b4-96bf-ec3d59133569-kube-api-access-kzgxs" (OuterVolumeSpecName: "kube-api-access-kzgxs") pod "b2edf8de-176a-44b4-96bf-ec3d59133569" (UID: "b2edf8de-176a-44b4-96bf-ec3d59133569"). InnerVolumeSpecName "kube-api-access-kzgxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.647904 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzgxs\" (UniqueName: \"kubernetes.io/projected/b2edf8de-176a-44b4-96bf-ec3d59133569-kube-api-access-kzgxs\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.647942 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.989298 4862 generic.go:334] "Generic (PLEG): container finished" podID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerID="f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e" exitCode=0 Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.989377 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s57qh" Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.989413 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s57qh" event={"ID":"b2edf8de-176a-44b4-96bf-ec3d59133569","Type":"ContainerDied","Data":"f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e"} Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.989653 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s57qh" event={"ID":"b2edf8de-176a-44b4-96bf-ec3d59133569","Type":"ContainerDied","Data":"b31d5a0b736b30a57a50dc8c6f37a2a26a5d141dd1ad43a3babd1edaacfc0083"} Nov 28 11:46:41 crc kubenswrapper[4862]: I1128 11:46:41.989684 4862 scope.go:117] "RemoveContainer" containerID="f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.014722 4862 scope.go:117] "RemoveContainer" containerID="65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.046253 4862 scope.go:117] "RemoveContainer" containerID="643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.077678 4862 scope.go:117] "RemoveContainer" containerID="f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e" Nov 28 11:46:42 crc kubenswrapper[4862]: E1128 11:46:42.078061 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e\": container with ID starting with f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e not found: ID does not exist" containerID="f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.078124 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e"} err="failed to get container status \"f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e\": rpc error: code = NotFound desc = could not find container \"f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e\": container with ID starting with f24088251246daa05caa606ea2c4bf85e2924ce8178e3bff7bbaf057aaabde5e not found: ID does not exist" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.078152 4862 scope.go:117] "RemoveContainer" containerID="65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5" Nov 28 11:46:42 crc kubenswrapper[4862]: E1128 11:46:42.078479 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5\": container with ID starting with 65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5 not found: ID does not exist" containerID="65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.078511 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5"} err="failed to get container status \"65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5\": rpc error: code = NotFound desc = could not find container \"65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5\": container with ID starting with 65adeec6d19c2af347f3ab22168efa094dad8f9f271384035af3cd7f1270ede5 not found: ID does not exist" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.078546 4862 scope.go:117] "RemoveContainer" containerID="643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3" Nov 28 11:46:42 crc kubenswrapper[4862]: E1128 11:46:42.079072 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3\": container with ID starting with 643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3 not found: ID does not exist" containerID="643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.079119 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3"} err="failed to get container status \"643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3\": rpc error: code = NotFound desc = could not find container \"643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3\": container with ID starting with 643b4ac76c535f617f1c3799d1edefe6d1bc97826761a714fff1bbb04aec51f3 not found: ID does not exist" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.722137 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2edf8de-176a-44b4-96bf-ec3d59133569" (UID: "b2edf8de-176a-44b4-96bf-ec3d59133569"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.770200 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2edf8de-176a-44b4-96bf-ec3d59133569-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.923220 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s57qh"] Nov 28 11:46:42 crc kubenswrapper[4862]: I1128 11:46:42.938444 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s57qh"] Nov 28 11:46:43 crc kubenswrapper[4862]: I1128 11:46:43.247453 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 11:46:43 crc kubenswrapper[4862]: I1128 11:46:43.247516 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 11:46:43 crc kubenswrapper[4862]: I1128 11:46:43.369281 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 11:46:44 crc kubenswrapper[4862]: I1128 11:46:44.267363 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:44 crc kubenswrapper[4862]: I1128 11:46:44.267391 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:44 crc kubenswrapper[4862]: I1128 11:46:44.863628 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2edf8de-176a-44b4-96bf-ec3d59133569" path="/var/lib/kubelet/pods/b2edf8de-176a-44b4-96bf-ec3d59133569/volumes" Nov 28 11:46:46 crc kubenswrapper[4862]: I1128 11:46:46.341358 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 11:46:46 crc kubenswrapper[4862]: I1128 11:46:46.343955 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 11:46:47 crc kubenswrapper[4862]: I1128 11:46:47.360398 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:47 crc kubenswrapper[4862]: I1128 11:46:47.360517 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 11:46:48 crc kubenswrapper[4862]: I1128 11:46:48.369240 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 11:46:48 crc kubenswrapper[4862]: I1128 11:46:48.402824 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 11:46:49 crc kubenswrapper[4862]: I1128 11:46:49.115665 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 11:46:53 crc kubenswrapper[4862]: I1128 11:46:53.255938 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 11:46:53 crc kubenswrapper[4862]: I1128 11:46:53.257070 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 11:46:53 crc kubenswrapper[4862]: I1128 11:46:53.261783 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 11:46:53 crc kubenswrapper[4862]: I1128 11:46:53.281600 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 11:46:54 crc kubenswrapper[4862]: I1128 11:46:54.136490 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 11:46:54 crc kubenswrapper[4862]: I1128 11:46:54.147477 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 11:46:55 crc kubenswrapper[4862]: I1128 11:46:55.102363 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 11:46:56 crc kubenswrapper[4862]: I1128 11:46:56.352116 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 11:46:56 crc kubenswrapper[4862]: I1128 11:46:56.352540 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 11:46:56 crc kubenswrapper[4862]: I1128 11:46:56.362324 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 11:46:56 crc kubenswrapper[4862]: I1128 11:46:56.363024 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 11:47:04 crc kubenswrapper[4862]: I1128 11:47:04.883921 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h5tks"] Nov 28 11:47:04 crc kubenswrapper[4862]: E1128 11:47:04.885313 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerName="registry-server" Nov 28 11:47:04 crc kubenswrapper[4862]: I1128 11:47:04.885340 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerName="registry-server" Nov 28 11:47:04 crc kubenswrapper[4862]: E1128 11:47:04.885389 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerName="extract-content" Nov 28 11:47:04 crc kubenswrapper[4862]: I1128 11:47:04.885403 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerName="extract-content" Nov 28 11:47:04 crc kubenswrapper[4862]: E1128 11:47:04.885423 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerName="extract-utilities" Nov 28 11:47:04 crc kubenswrapper[4862]: I1128 11:47:04.885437 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerName="extract-utilities" Nov 28 11:47:04 crc kubenswrapper[4862]: I1128 11:47:04.885819 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2edf8de-176a-44b4-96bf-ec3d59133569" containerName="registry-server" Nov 28 11:47:04 crc kubenswrapper[4862]: I1128 11:47:04.888499 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:04 crc kubenswrapper[4862]: I1128 11:47:04.907706 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5tks"] Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.025940 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r64mh\" (UniqueName: \"kubernetes.io/projected/8feb55d1-4c01-4321-8f7e-0723aef07525-kube-api-access-r64mh\") pod \"redhat-marketplace-h5tks\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.025985 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-utilities\") pod \"redhat-marketplace-h5tks\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.026019 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-catalog-content\") pod \"redhat-marketplace-h5tks\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.128121 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r64mh\" (UniqueName: \"kubernetes.io/projected/8feb55d1-4c01-4321-8f7e-0723aef07525-kube-api-access-r64mh\") pod \"redhat-marketplace-h5tks\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.128439 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-utilities\") pod \"redhat-marketplace-h5tks\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.128475 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-catalog-content\") pod \"redhat-marketplace-h5tks\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.129014 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-catalog-content\") pod \"redhat-marketplace-h5tks\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.129154 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-utilities\") pod \"redhat-marketplace-h5tks\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.154182 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r64mh\" (UniqueName: \"kubernetes.io/projected/8feb55d1-4c01-4321-8f7e-0723aef07525-kube-api-access-r64mh\") pod \"redhat-marketplace-h5tks\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.212374 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:05 crc kubenswrapper[4862]: I1128 11:47:05.712592 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5tks"] Nov 28 11:47:05 crc kubenswrapper[4862]: W1128 11:47:05.725551 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8feb55d1_4c01_4321_8f7e_0723aef07525.slice/crio-020f756d4f01ce79184a5e315e7e03d7985132f103da20ab927d85b1905a9012 WatchSource:0}: Error finding container 020f756d4f01ce79184a5e315e7e03d7985132f103da20ab927d85b1905a9012: Status 404 returned error can't find the container with id 020f756d4f01ce79184a5e315e7e03d7985132f103da20ab927d85b1905a9012 Nov 28 11:47:06 crc kubenswrapper[4862]: I1128 11:47:06.316626 4862 generic.go:334] "Generic (PLEG): container finished" podID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerID="d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97" exitCode=0 Nov 28 11:47:06 crc kubenswrapper[4862]: I1128 11:47:06.316691 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5tks" event={"ID":"8feb55d1-4c01-4321-8f7e-0723aef07525","Type":"ContainerDied","Data":"d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97"} Nov 28 11:47:06 crc kubenswrapper[4862]: I1128 11:47:06.317053 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5tks" event={"ID":"8feb55d1-4c01-4321-8f7e-0723aef07525","Type":"ContainerStarted","Data":"020f756d4f01ce79184a5e315e7e03d7985132f103da20ab927d85b1905a9012"} Nov 28 11:47:07 crc kubenswrapper[4862]: I1128 11:47:07.331533 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5tks" event={"ID":"8feb55d1-4c01-4321-8f7e-0723aef07525","Type":"ContainerStarted","Data":"2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8"} Nov 28 11:47:08 crc kubenswrapper[4862]: I1128 11:47:08.350201 4862 generic.go:334] "Generic (PLEG): container finished" podID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerID="2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8" exitCode=0 Nov 28 11:47:08 crc kubenswrapper[4862]: I1128 11:47:08.350261 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5tks" event={"ID":"8feb55d1-4c01-4321-8f7e-0723aef07525","Type":"ContainerDied","Data":"2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8"} Nov 28 11:47:09 crc kubenswrapper[4862]: I1128 11:47:09.367333 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5tks" event={"ID":"8feb55d1-4c01-4321-8f7e-0723aef07525","Type":"ContainerStarted","Data":"870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548"} Nov 28 11:47:09 crc kubenswrapper[4862]: I1128 11:47:09.399671 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h5tks" podStartSLOduration=2.821796608 podStartE2EDuration="5.399653193s" podCreationTimestamp="2025-11-28 11:47:04 +0000 UTC" firstStartedPulling="2025-11-28 11:47:06.318929271 +0000 UTC m=+1503.851443202" lastFinishedPulling="2025-11-28 11:47:08.896785826 +0000 UTC m=+1506.429299787" observedRunningTime="2025-11-28 11:47:09.397121341 +0000 UTC m=+1506.929635292" watchObservedRunningTime="2025-11-28 11:47:09.399653193 +0000 UTC m=+1506.932167114" Nov 28 11:47:15 crc kubenswrapper[4862]: I1128 11:47:15.213239 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:15 crc kubenswrapper[4862]: I1128 11:47:15.213770 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:15 crc kubenswrapper[4862]: I1128 11:47:15.275625 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:15 crc kubenswrapper[4862]: I1128 11:47:15.482348 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:15 crc kubenswrapper[4862]: I1128 11:47:15.528890 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5tks"] Nov 28 11:47:17 crc kubenswrapper[4862]: I1128 11:47:17.468460 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h5tks" podUID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerName="registry-server" containerID="cri-o://870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548" gracePeriod=2 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.130879 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.131394 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerName="cinder-scheduler" containerID="cri-o://8031be805d28d00e73f2d74cac7b5638f7feeaa4fd5d9e786af7a60f92caf4cc" gracePeriod=30 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.131814 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerName="probe" containerID="cri-o://5d77e5ca098b36fcc98d53cfa5e572df42e3b1667b2235d6c42ff31fe0c66c1f" gracePeriod=30 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.148406 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.148634 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="23891b0e-3fc7-4d2f-9abf-96b57b09ca01" containerName="openstackclient" containerID="cri-o://36eb5af2c95a5dd49963c18b778f72ccbf03d09076fbe15508a1df0b8532d372" gracePeriod=2 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.180152 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.199209 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.255434 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.255666 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api-log" containerID="cri-o://72dd7e05b0e98629a3547aa5ced1d209fea4cbcd7cb7dc51836836e79ae4b826" gracePeriod=30 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.255781 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api" containerID="cri-o://2732ee7a075e6578c8913e90a93812b8f69feb263ac51f3983a02e4f5cb85ed0" gracePeriod=30 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.271742 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.321063 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.321296 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data podName:88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:18.821278296 +0000 UTC m=+1516.353792207 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data") pod "rabbitmq-server-0" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356") : configmap "rabbitmq-config-data" not found Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.421934 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-catalog-content\") pod \"8feb55d1-4c01-4321-8f7e-0723aef07525\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.422201 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r64mh\" (UniqueName: \"kubernetes.io/projected/8feb55d1-4c01-4321-8f7e-0723aef07525-kube-api-access-r64mh\") pod \"8feb55d1-4c01-4321-8f7e-0723aef07525\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.422226 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-utilities\") pod \"8feb55d1-4c01-4321-8f7e-0723aef07525\" (UID: \"8feb55d1-4c01-4321-8f7e-0723aef07525\") " Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.431080 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-utilities" (OuterVolumeSpecName: "utilities") pod "8feb55d1-4c01-4321-8f7e-0723aef07525" (UID: "8feb55d1-4c01-4321-8f7e-0723aef07525"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.431431 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8feb55d1-4c01-4321-8f7e-0723aef07525-kube-api-access-r64mh" (OuterVolumeSpecName: "kube-api-access-r64mh") pod "8feb55d1-4c01-4321-8f7e-0723aef07525" (UID: "8feb55d1-4c01-4321-8f7e-0723aef07525"). InnerVolumeSpecName "kube-api-access-r64mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.474425 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.482647 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8feb55d1-4c01-4321-8f7e-0723aef07525" (UID: "8feb55d1-4c01-4321-8f7e-0723aef07525"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.493462 4862 generic.go:334] "Generic (PLEG): container finished" podID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerID="870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548" exitCode=0 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.493580 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5tks" event={"ID":"8feb55d1-4c01-4321-8f7e-0723aef07525","Type":"ContainerDied","Data":"870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548"} Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.493609 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5tks" event={"ID":"8feb55d1-4c01-4321-8f7e-0723aef07525","Type":"ContainerDied","Data":"020f756d4f01ce79184a5e315e7e03d7985132f103da20ab927d85b1905a9012"} Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.493626 4862 scope.go:117] "RemoveContainer" containerID="870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.493723 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5tks" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.524118 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r64mh\" (UniqueName: \"kubernetes.io/projected/8feb55d1-4c01-4321-8f7e-0723aef07525-kube-api-access-r64mh\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.524146 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.524155 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8feb55d1-4c01-4321-8f7e-0723aef07525-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.555574 4862 generic.go:334] "Generic (PLEG): container finished" podID="ffc2928a-7858-4117-8786-723965b126f4" containerID="72dd7e05b0e98629a3547aa5ced1d209fea4cbcd7cb7dc51836836e79ae4b826" exitCode=143 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.555625 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ffc2928a-7858-4117-8786-723965b126f4","Type":"ContainerDied","Data":"72dd7e05b0e98629a3547aa5ced1d209fea4cbcd7cb7dc51836836e79ae4b826"} Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.569393 4862 scope.go:117] "RemoveContainer" containerID="2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.590150 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican8d34-account-delete-bdd2b"] Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.595352 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerName="extract-utilities" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.595384 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerName="extract-utilities" Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.595405 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerName="registry-server" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.595411 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerName="registry-server" Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.595428 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23891b0e-3fc7-4d2f-9abf-96b57b09ca01" containerName="openstackclient" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.595434 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="23891b0e-3fc7-4d2f-9abf-96b57b09ca01" containerName="openstackclient" Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.595451 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerName="extract-content" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.595458 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerName="extract-content" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.595716 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="23891b0e-3fc7-4d2f-9abf-96b57b09ca01" containerName="openstackclient" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.595740 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="8feb55d1-4c01-4321-8f7e-0723aef07525" containerName="registry-server" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.596379 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.633465 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5tks"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.641221 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5tks"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.656293 4862 scope.go:117] "RemoveContainer" containerID="d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.668153 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0c5d7-account-delete-tw6mt"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.677314 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.680233 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.680734 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="69897166-b5c9-404e-ac62-0b55f0960278" containerName="openstack-network-exporter" containerID="cri-o://670f94a7e5c8ef75b6abff70cce614bc7b826f6dcb5baf65b275f5432887d508" gracePeriod=300 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.688646 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican8d34-account-delete-bdd2b"] Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.729034 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.729111 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data podName:527b2bce-1186-4186-8992-a3a63d2d2b22 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:19.229076729 +0000 UTC m=+1516.761590730 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data") pod "rabbitmq-cell1-server-0" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22") : configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.759272 4862 scope.go:117] "RemoveContainer" containerID="870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.830493 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts\") pod \"barbican8d34-account-delete-bdd2b\" (UID: \"ed3d880d-6762-4d57-881d-bc585318d500\") " pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.830550 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzsw9\" (UniqueName: \"kubernetes.io/projected/75f196ee-9a0c-414f-a629-5b270c412c24-kube-api-access-lzsw9\") pod \"novacell0c5d7-account-delete-tw6mt\" (UID: \"75f196ee-9a0c-414f-a629-5b270c412c24\") " pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.830596 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxn4v\" (UniqueName: \"kubernetes.io/projected/ed3d880d-6762-4d57-881d-bc585318d500-kube-api-access-jxn4v\") pod \"barbican8d34-account-delete-bdd2b\" (UID: \"ed3d880d-6762-4d57-881d-bc585318d500\") " pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.830745 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75f196ee-9a0c-414f-a629-5b270c412c24-operator-scripts\") pod \"novacell0c5d7-account-delete-tw6mt\" (UID: \"75f196ee-9a0c-414f-a629-5b270c412c24\") " pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.830953 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.831003 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data podName:88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:19.830985518 +0000 UTC m=+1517.363499439 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data") pod "rabbitmq-server-0" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356") : configmap "rabbitmq-config-data" not found Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.844988 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548\": container with ID starting with 870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548 not found: ID does not exist" containerID="870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.845032 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548"} err="failed to get container status \"870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548\": rpc error: code = NotFound desc = could not find container \"870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548\": container with ID starting with 870322777e5277980ebb10f7ffbacb903853c94069627e2168fb5942a5e87548 not found: ID does not exist" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.845058 4862 scope.go:117] "RemoveContainer" containerID="2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8" Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.846745 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8\": container with ID starting with 2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8 not found: ID does not exist" containerID="2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.846799 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8"} err="failed to get container status \"2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8\": rpc error: code = NotFound desc = could not find container \"2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8\": container with ID starting with 2fcfd57e723b0681c53118b47b8edd7612a9f1dabbae2f5a6bc5761c165ce2f8 not found: ID does not exist" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.846827 4862 scope.go:117] "RemoveContainer" containerID="d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97" Nov 28 11:47:18 crc kubenswrapper[4862]: E1128 11:47:18.860523 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97\": container with ID starting with d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97 not found: ID does not exist" containerID="d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.860787 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97"} err="failed to get container status \"d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97\": rpc error: code = NotFound desc = could not find container \"d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97\": container with ID starting with d7676aabaff8965664a40198778c4a8d70c08c7aad732023646d3522fee6dd97 not found: ID does not exist" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.911248 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="69897166-b5c9-404e-ac62-0b55f0960278" containerName="ovsdbserver-sb" containerID="cri-o://a1a8cae69793fcdeda0115c040ea95de4e540df38f817fe5d2600245f84cba4b" gracePeriod=300 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.933257 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75f196ee-9a0c-414f-a629-5b270c412c24-operator-scripts\") pod \"novacell0c5d7-account-delete-tw6mt\" (UID: \"75f196ee-9a0c-414f-a629-5b270c412c24\") " pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.933339 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts\") pod \"barbican8d34-account-delete-bdd2b\" (UID: \"ed3d880d-6762-4d57-881d-bc585318d500\") " pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.933363 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzsw9\" (UniqueName: \"kubernetes.io/projected/75f196ee-9a0c-414f-a629-5b270c412c24-kube-api-access-lzsw9\") pod \"novacell0c5d7-account-delete-tw6mt\" (UID: \"75f196ee-9a0c-414f-a629-5b270c412c24\") " pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.933395 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxn4v\" (UniqueName: \"kubernetes.io/projected/ed3d880d-6762-4d57-881d-bc585318d500-kube-api-access-jxn4v\") pod \"barbican8d34-account-delete-bdd2b\" (UID: \"ed3d880d-6762-4d57-881d-bc585318d500\") " pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.935001 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75f196ee-9a0c-414f-a629-5b270c412c24-operator-scripts\") pod \"novacell0c5d7-account-delete-tw6mt\" (UID: \"75f196ee-9a0c-414f-a629-5b270c412c24\") " pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.936036 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts\") pod \"barbican8d34-account-delete-bdd2b\" (UID: \"ed3d880d-6762-4d57-881d-bc585318d500\") " pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.950118 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8feb55d1-4c01-4321-8f7e-0723aef07525" path="/var/lib/kubelet/pods/8feb55d1-4c01-4321-8f7e-0723aef07525/volumes" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.950779 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0c5d7-account-delete-tw6mt"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.957203 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.957442 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="ovn-northd" containerID="cri-o://3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6" gracePeriod=30 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.957561 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="openstack-network-exporter" containerID="cri-o://23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3" gracePeriod=30 Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.963739 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi144c-account-delete-ls9c4"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.964934 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.982847 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi144c-account-delete-ls9c4"] Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.992848 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzsw9\" (UniqueName: \"kubernetes.io/projected/75f196ee-9a0c-414f-a629-5b270c412c24-kube-api-access-lzsw9\") pod \"novacell0c5d7-account-delete-tw6mt\" (UID: \"75f196ee-9a0c-414f-a629-5b270c412c24\") " pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:18 crc kubenswrapper[4862]: I1128 11:47:18.995598 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxn4v\" (UniqueName: \"kubernetes.io/projected/ed3d880d-6762-4d57-881d-bc585318d500-kube-api-access-jxn4v\") pod \"barbican8d34-account-delete-bdd2b\" (UID: \"ed3d880d-6762-4d57-881d-bc585318d500\") " pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.022695 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placemente4ca-account-delete-hcmct"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.023928 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.037417 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qqsr\" (UniqueName: \"kubernetes.io/projected/409c7014-982c-4c60-b374-0ff46c398db4-kube-api-access-2qqsr\") pod \"novaapi144c-account-delete-ls9c4\" (UID: \"409c7014-982c-4c60-b374-0ff46c398db4\") " pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.037565 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts\") pod \"novaapi144c-account-delete-ls9c4\" (UID: \"409c7014-982c-4c60-b374-0ff46c398db4\") " pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.043931 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.099541 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placemente4ca-account-delete-hcmct"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.140272 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts\") pod \"novaapi144c-account-delete-ls9c4\" (UID: \"409c7014-982c-4c60-b374-0ff46c398db4\") " pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.140402 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jznv\" (UniqueName: \"kubernetes.io/projected/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-kube-api-access-9jznv\") pod \"placemente4ca-account-delete-hcmct\" (UID: \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\") " pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.140434 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qqsr\" (UniqueName: \"kubernetes.io/projected/409c7014-982c-4c60-b374-0ff46c398db4-kube-api-access-2qqsr\") pod \"novaapi144c-account-delete-ls9c4\" (UID: \"409c7014-982c-4c60-b374-0ff46c398db4\") " pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.140513 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-operator-scripts\") pod \"placemente4ca-account-delete-hcmct\" (UID: \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\") " pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.143543 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="ovn-northd" probeResult="failure" output=< Nov 28 11:47:19 crc kubenswrapper[4862]: 2025-11-28T11:47:19Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Nov 28 11:47:19 crc kubenswrapper[4862]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Nov 28 11:47:19 crc kubenswrapper[4862]: > Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.148081 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts\") pod \"novaapi144c-account-delete-ls9c4\" (UID: \"409c7014-982c-4c60-b374-0ff46c398db4\") " pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.209749 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-t9gw8"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.249250 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jznv\" (UniqueName: \"kubernetes.io/projected/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-kube-api-access-9jznv\") pod \"placemente4ca-account-delete-hcmct\" (UID: \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\") " pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:19 crc kubenswrapper[4862]: E1128 11:47:19.251603 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:19 crc kubenswrapper[4862]: E1128 11:47:19.256627 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data podName:527b2bce-1186-4186-8992-a3a63d2d2b22 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:20.256598549 +0000 UTC m=+1517.789112470 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data") pod "rabbitmq-cell1-server-0" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22") : configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.261303 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-operator-scripts\") pod \"placemente4ca-account-delete-hcmct\" (UID: \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\") " pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.262241 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-operator-scripts\") pod \"placemente4ca-account-delete-hcmct\" (UID: \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\") " pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.264959 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.266685 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qqsr\" (UniqueName: \"kubernetes.io/projected/409c7014-982c-4c60-b374-0ff46c398db4-kube-api-access-2qqsr\") pod \"novaapi144c-account-delete-ls9c4\" (UID: \"409c7014-982c-4c60-b374-0ff46c398db4\") " pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.331979 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jznv\" (UniqueName: \"kubernetes.io/projected/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-kube-api-access-9jznv\") pod \"placemente4ca-account-delete-hcmct\" (UID: \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\") " pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.334163 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dnh4p"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.371913 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.427445 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.444837 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-t9gw8"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.482513 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="ovn-northd" probeResult="failure" output=< Nov 28 11:47:19 crc kubenswrapper[4862]: 2025-11-28T11:47:19Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Nov 28 11:47:19 crc kubenswrapper[4862]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Nov 28 11:47:19 crc kubenswrapper[4862]: 2025-11-28T11:47:19Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Nov 28 11:47:19 crc kubenswrapper[4862]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Nov 28 11:47:19 crc kubenswrapper[4862]: > Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.510149 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-bnflp"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.556743 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dnh4p"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.577466 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-bnflp"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.625278 4862 generic.go:334] "Generic (PLEG): container finished" podID="69897166-b5c9-404e-ac62-0b55f0960278" containerID="670f94a7e5c8ef75b6abff70cce614bc7b826f6dcb5baf65b275f5432887d508" exitCode=2 Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.625404 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69897166-b5c9-404e-ac62-0b55f0960278","Type":"ContainerDied","Data":"670f94a7e5c8ef75b6abff70cce614bc7b826f6dcb5baf65b275f5432887d508"} Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.636610 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutronc4cb-account-delete-9jrbn"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.646501 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.674893 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronc4cb-account-delete-9jrbn"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.688050 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j44b\" (UniqueName: \"kubernetes.io/projected/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-kube-api-access-8j44b\") pod \"neutronc4cb-account-delete-9jrbn\" (UID: \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\") " pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.688471 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts\") pod \"neutronc4cb-account-delete-9jrbn\" (UID: \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\") " pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.724220 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-5lsz4"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.749786 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-5lsz4"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.771364 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hkgwr"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.797706 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j44b\" (UniqueName: \"kubernetes.io/projected/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-kube-api-access-8j44b\") pod \"neutronc4cb-account-delete-9jrbn\" (UID: \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\") " pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.798238 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts\") pod \"neutronc4cb-account-delete-9jrbn\" (UID: \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\") " pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.799085 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts\") pod \"neutronc4cb-account-delete-9jrbn\" (UID: \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\") " pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.802722 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-hdvv6"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.845288 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j44b\" (UniqueName: \"kubernetes.io/projected/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-kube-api-access-8j44b\") pod \"neutronc4cb-account-delete-9jrbn\" (UID: \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\") " pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.849654 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-flbld"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.849883 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-flbld" podUID="53d62b74-0406-47fe-8a9a-36cc6f288f27" containerName="openstack-network-exporter" containerID="cri-o://a1e1c2cca57dec29a126aac011c043be5314d2cdd1005a224a53029e52b33e16" gracePeriod=30 Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.893885 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-47sh5"] Nov 28 11:47:19 crc kubenswrapper[4862]: E1128 11:47:19.912477 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 11:47:19 crc kubenswrapper[4862]: E1128 11:47:19.912541 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data podName:88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:21.912520535 +0000 UTC m=+1519.445034516 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data") pod "rabbitmq-server-0" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356") : configmap "rabbitmq-config-data" not found Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.953179 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-47sh5"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.978503 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance056c-account-delete-cn6wm"] Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.984709 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:19 crc kubenswrapper[4862]: I1128 11:47:19.991918 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance056c-account-delete-cn6wm"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.007252 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.008223 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerName="openstack-network-exporter" containerID="cri-o://e91e1d7ad1e2a8876f0520009135a4b0b1915dd5112109b61fca65064cefcd34" gracePeriod=300 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.019444 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc646c8f9-jmhdw"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.019693 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" podUID="106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" containerName="dnsmasq-dns" containerID="cri-o://d351e93457fe5a70f380ffce1f54e4bea632ff6f5419983422bc2fc2947f982d" gracePeriod=10 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.035546 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-qltsz"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.045554 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-qltsz"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.055500 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-5jcj5"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.087950 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.111272 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zqg7\" (UniqueName: \"kubernetes.io/projected/71c0d0c9-ed19-41a0-a3d1-92088179bf04-kube-api-access-4zqg7\") pod \"glance056c-account-delete-cn6wm\" (UID: \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\") " pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.111403 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-5jcj5"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.115055 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71c0d0c9-ed19-41a0-a3d1-92088179bf04-operator-scripts\") pod \"glance056c-account-delete-cn6wm\" (UID: \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\") " pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.135165 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder68e8-account-delete-rv4w4"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.136932 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.140042 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerName="ovsdbserver-nb" containerID="cri-o://ee8d7e9a14861e17f472a1f44420408da6cefecb358b29bd4e8a5166746780c2" gracePeriod=300 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.159024 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder68e8-account-delete-rv4w4"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.200927 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-xchsv"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.220105 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zqg7\" (UniqueName: \"kubernetes.io/projected/71c0d0c9-ed19-41a0-a3d1-92088179bf04-kube-api-access-4zqg7\") pod \"glance056c-account-delete-cn6wm\" (UID: \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\") " pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.220156 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts\") pod \"cinder68e8-account-delete-rv4w4\" (UID: \"880d930f-ea3a-4a7b-a437-5b164fbea09e\") " pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.220290 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwb7s\" (UniqueName: \"kubernetes.io/projected/880d930f-ea3a-4a7b-a437-5b164fbea09e-kube-api-access-fwb7s\") pod \"cinder68e8-account-delete-rv4w4\" (UID: \"880d930f-ea3a-4a7b-a437-5b164fbea09e\") " pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.220333 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71c0d0c9-ed19-41a0-a3d1-92088179bf04-operator-scripts\") pod \"glance056c-account-delete-cn6wm\" (UID: \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\") " pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.222582 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71c0d0c9-ed19-41a0-a3d1-92088179bf04-operator-scripts\") pod \"glance056c-account-delete-cn6wm\" (UID: \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\") " pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.246881 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-xchsv"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.261680 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.263862 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zqg7\" (UniqueName: \"kubernetes.io/projected/71c0d0c9-ed19-41a0-a3d1-92088179bf04-kube-api-access-4zqg7\") pod \"glance056c-account-delete-cn6wm\" (UID: \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\") " pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.274731 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275227 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-server" containerID="cri-o://540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275632 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="swift-recon-cron" containerID="cri-o://c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275683 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-auditor" containerID="cri-o://df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275715 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-replicator" containerID="cri-o://d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275747 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-server" containerID="cri-o://563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275728 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-updater" containerID="cri-o://f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275770 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="rsync" containerID="cri-o://775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275803 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-reaper" containerID="cri-o://53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275835 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-auditor" containerID="cri-o://aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275870 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-replicator" containerID="cri-o://b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275903 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-expirer" containerID="cri-o://0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275956 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-updater" containerID="cri-o://84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.275998 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-auditor" containerID="cri-o://2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.276040 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-replicator" containerID="cri-o://891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.276079 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-server" containerID="cri-o://4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.283833 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-69477d9ff4-9mk7g"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.284133 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-69477d9ff4-9mk7g" podUID="ca62671f-2a31-4980-93d9-2239defbfa57" containerName="placement-log" containerID="cri-o://ec72be9e8e9e4302a004ec5624c9e7d3c1b7137a5f5cc1b75511dcf0b235953b" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.286587 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-69477d9ff4-9mk7g" podUID="ca62671f-2a31-4980-93d9-2239defbfa57" containerName="placement-api" containerID="cri-o://a7c34e84cffc71a4ccf07325940e732e6d5ffc06d972e6e93c18df3db794e0b9" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.299288 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.299575 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-log" containerID="cri-o://a303ac97d092846a039471dffa93b8705481508ae03493b999176d9423f8a8c4" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.299729 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-metadata" containerID="cri-o://15217408eaf0c5ecfdbf05adb57c81819bf35870f42f21b0ecf1bd8c4c5ff8a8" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.329612 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: E1128 11:47:20.331285 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:20 crc kubenswrapper[4862]: E1128 11:47:20.331350 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data podName:527b2bce-1186-4186-8992-a3a63d2d2b22 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:22.331332492 +0000 UTC m=+1519.863846413 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data") pod "rabbitmq-cell1-server-0" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22") : configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.333245 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwb7s\" (UniqueName: \"kubernetes.io/projected/880d930f-ea3a-4a7b-a437-5b164fbea09e-kube-api-access-fwb7s\") pod \"cinder68e8-account-delete-rv4w4\" (UID: \"880d930f-ea3a-4a7b-a437-5b164fbea09e\") " pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.333588 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts\") pod \"cinder68e8-account-delete-rv4w4\" (UID: \"880d930f-ea3a-4a7b-a437-5b164fbea09e\") " pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.334647 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts\") pod \"cinder68e8-account-delete-rv4w4\" (UID: \"880d930f-ea3a-4a7b-a437-5b164fbea09e\") " pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.383261 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-srt25"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.386889 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.393448 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwb7s\" (UniqueName: \"kubernetes.io/projected/880d930f-ea3a-4a7b-a437-5b164fbea09e-kube-api-access-fwb7s\") pod \"cinder68e8-account-delete-rv4w4\" (UID: \"880d930f-ea3a-4a7b-a437-5b164fbea09e\") " pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.405704 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-srt25"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.422141 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6dc889566f-77hs4"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.428595 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6dc889566f-77hs4" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerName="neutron-api" containerID="cri-o://9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.429022 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6dc889566f-77hs4" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerName="neutron-httpd" containerID="cri-o://76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.433688 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" containerName="rabbitmq" containerID="cri-o://dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f" gracePeriod=604800 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.441044 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.441228 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a824cc2c-2812-4d91-9e9f-9736a27fe794" containerName="nova-scheduler-scheduler" containerID="cri-o://0682b22558ac57d011862a3e8a94d6b4334e16c0f890d30fb6ae86e9faaac01e" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.451221 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-c990-account-create-update-2sxzh"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.475752 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-c990-account-create-update-2sxzh"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.497145 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.497430 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-log" containerID="cri-o://adb196b1aa7dfec6e01189bb208acacebcbed93cfaed7fde33e15993b31a31f9" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.497930 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-api" containerID="cri-o://69388439d162f0fc46186cc6f4708492782333d0c2f8c07c9238c3381aaf99ad" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.504664 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.504891 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerName="glance-log" containerID="cri-o://236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.505024 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerName="glance-httpd" containerID="cri-o://ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.512213 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0c5d7-account-delete-tw6mt"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.521008 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6df5c754df-6lvmr"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.521324 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6df5c754df-6lvmr" podUID="d5203b21-897c-4331-a878-1d5c949a3e75" containerName="barbican-worker-log" containerID="cri-o://cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.521747 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6df5c754df-6lvmr" podUID="d5203b21-897c-4331-a878-1d5c949a3e75" containerName="barbican-worker" containerID="cri-o://31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.531267 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-689b7bfc65-h98mj"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.531495 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-689b7bfc65-h98mj" podUID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerName="proxy-httpd" containerID="cri-o://b66e34c72b94210fa8e1b94676bc9d86fdb52829035fa8a84ba5bbf0ebf630c5" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.531603 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-689b7bfc65-h98mj" podUID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerName="proxy-server" containerID="cri-o://cb861c743ce00570c223afefb40f3cc5ed244b9682a5c2e0bbf669dba92ca34f" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.538847 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.539064 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerName="glance-log" containerID="cri-o://7e91caadb693c64c084738db969219970babbd935caf36c5d56c0b2e9a460f4a" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.539504 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerName="glance-httpd" containerID="cri-o://b622cbdd5687d59d52eeed34f730456b8b029c86f0df4d63c0373c42dc4992e7" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.545878 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7756445dd6-vwrsv"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.546116 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" podUID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerName="barbican-keystone-listener-log" containerID="cri-o://d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.546412 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" podUID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerName="barbican-keystone-listener" containerID="cri-o://b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.553243 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-857784bb5b-4pc8q"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.553456 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-857784bb5b-4pc8q" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api-log" containerID="cri-o://39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.553934 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-857784bb5b-4pc8q" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api" containerID="cri-o://773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.571042 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.571301 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="ae02e889-f98f-4678-b15f-d91f711083f9" containerName="nova-cell1-conductor-conductor" containerID="cri-o://e487aac69ef596761ce6caee885f81d9e334146dc7a0a9ba8b24df61e5dc1e8f" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.581478 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f8krc"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.594143 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f8krc"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.608569 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.612138 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2tm6f"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.618830 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.619263 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="01f7b6ff-ee76-4961-9baf-930743da6143" containerName="nova-cell0-conductor-conductor" containerID="cri-o://821e6798c6600a5f98070c95203ea984a4b500e28b5b89da8f188ebceafa93c7" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.627588 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.627833 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="73c57ace-65f0-40ab-996a-e5595851177a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e4fbe66671af80c3245a6da3f4c0fcf1f2a7319780d1ab11a9026ea78828037b" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.635429 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2tm6f"] Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.685397 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0c5d7-account-delete-tw6mt" event={"ID":"75f196ee-9a0c-414f-a629-5b270c412c24","Type":"ContainerStarted","Data":"ac599f52e26cdf7e02a4c590fcf1a44aa33c68d4a7f0ba3b99172dad869beaf0"} Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.700584 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b/ovsdbserver-nb/0.log" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.700631 4862 generic.go:334] "Generic (PLEG): container finished" podID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerID="e91e1d7ad1e2a8876f0520009135a4b0b1915dd5112109b61fca65064cefcd34" exitCode=2 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.700648 4862 generic.go:334] "Generic (PLEG): container finished" podID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerID="ee8d7e9a14861e17f472a1f44420408da6cefecb358b29bd4e8a5166746780c2" exitCode=143 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.700734 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b","Type":"ContainerDied","Data":"e91e1d7ad1e2a8876f0520009135a4b0b1915dd5112109b61fca65064cefcd34"} Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.700760 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b","Type":"ContainerDied","Data":"ee8d7e9a14861e17f472a1f44420408da6cefecb358b29bd4e8a5166746780c2"} Nov 28 11:47:20 crc kubenswrapper[4862]: E1128 11:47:20.739608 4862 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 28 11:47:20 crc kubenswrapper[4862]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 28 11:47:20 crc kubenswrapper[4862]: + source /usr/local/bin/container-scripts/functions Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNBridge=br-int Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNRemote=tcp:localhost:6642 Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNEncapType=geneve Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNAvailabilityZones= Nov 28 11:47:20 crc kubenswrapper[4862]: ++ EnableChassisAsGateway=true Nov 28 11:47:20 crc kubenswrapper[4862]: ++ PhysicalNetworks= Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNHostName= Nov 28 11:47:20 crc kubenswrapper[4862]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 28 11:47:20 crc kubenswrapper[4862]: ++ ovs_dir=/var/lib/openvswitch Nov 28 11:47:20 crc kubenswrapper[4862]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 28 11:47:20 crc kubenswrapper[4862]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 28 11:47:20 crc kubenswrapper[4862]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 11:47:20 crc kubenswrapper[4862]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 11:47:20 crc kubenswrapper[4862]: + sleep 0.5 Nov 28 11:47:20 crc kubenswrapper[4862]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 11:47:20 crc kubenswrapper[4862]: + cleanup_ovsdb_server_semaphore Nov 28 11:47:20 crc kubenswrapper[4862]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 11:47:20 crc kubenswrapper[4862]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 28 11:47:20 crc kubenswrapper[4862]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-hdvv6" message=< Nov 28 11:47:20 crc kubenswrapper[4862]: Exiting ovsdb-server (5) [ OK ] Nov 28 11:47:20 crc kubenswrapper[4862]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 28 11:47:20 crc kubenswrapper[4862]: + source /usr/local/bin/container-scripts/functions Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNBridge=br-int Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNRemote=tcp:localhost:6642 Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNEncapType=geneve Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNAvailabilityZones= Nov 28 11:47:20 crc kubenswrapper[4862]: ++ EnableChassisAsGateway=true Nov 28 11:47:20 crc kubenswrapper[4862]: ++ PhysicalNetworks= Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNHostName= Nov 28 11:47:20 crc kubenswrapper[4862]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 28 11:47:20 crc kubenswrapper[4862]: ++ ovs_dir=/var/lib/openvswitch Nov 28 11:47:20 crc kubenswrapper[4862]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 28 11:47:20 crc kubenswrapper[4862]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 28 11:47:20 crc kubenswrapper[4862]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 11:47:20 crc kubenswrapper[4862]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 11:47:20 crc kubenswrapper[4862]: + sleep 0.5 Nov 28 11:47:20 crc kubenswrapper[4862]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 11:47:20 crc kubenswrapper[4862]: + cleanup_ovsdb_server_semaphore Nov 28 11:47:20 crc kubenswrapper[4862]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 11:47:20 crc kubenswrapper[4862]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 28 11:47:20 crc kubenswrapper[4862]: > Nov 28 11:47:20 crc kubenswrapper[4862]: E1128 11:47:20.739678 4862 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 28 11:47:20 crc kubenswrapper[4862]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 28 11:47:20 crc kubenswrapper[4862]: + source /usr/local/bin/container-scripts/functions Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNBridge=br-int Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNRemote=tcp:localhost:6642 Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNEncapType=geneve Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNAvailabilityZones= Nov 28 11:47:20 crc kubenswrapper[4862]: ++ EnableChassisAsGateway=true Nov 28 11:47:20 crc kubenswrapper[4862]: ++ PhysicalNetworks= Nov 28 11:47:20 crc kubenswrapper[4862]: ++ OVNHostName= Nov 28 11:47:20 crc kubenswrapper[4862]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 28 11:47:20 crc kubenswrapper[4862]: ++ ovs_dir=/var/lib/openvswitch Nov 28 11:47:20 crc kubenswrapper[4862]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 28 11:47:20 crc kubenswrapper[4862]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 28 11:47:20 crc kubenswrapper[4862]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 11:47:20 crc kubenswrapper[4862]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 11:47:20 crc kubenswrapper[4862]: + sleep 0.5 Nov 28 11:47:20 crc kubenswrapper[4862]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 11:47:20 crc kubenswrapper[4862]: + cleanup_ovsdb_server_semaphore Nov 28 11:47:20 crc kubenswrapper[4862]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 11:47:20 crc kubenswrapper[4862]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 28 11:47:20 crc kubenswrapper[4862]: > pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" containerID="cri-o://49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.739718 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" containerID="cri-o://49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" gracePeriod=30 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.757644 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_69897166-b5c9-404e-ac62-0b55f0960278/ovsdbserver-sb/0.log" Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.757695 4862 generic.go:334] "Generic (PLEG): container finished" podID="69897166-b5c9-404e-ac62-0b55f0960278" containerID="a1a8cae69793fcdeda0115c040ea95de4e540df38f817fe5d2600245f84cba4b" exitCode=143 Nov 28 11:47:20 crc kubenswrapper[4862]: I1128 11:47:20.757797 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69897166-b5c9-404e-ac62-0b55f0960278","Type":"ContainerDied","Data":"a1a8cae69793fcdeda0115c040ea95de4e540df38f817fe5d2600245f84cba4b"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.784571 4862 generic.go:334] "Generic (PLEG): container finished" podID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerID="236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb" exitCode=143 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.784878 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e8ef59ee-556c-4735-8cdf-e140aa608f86","Type":"ContainerDied","Data":"236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.793700 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="527b2bce-1186-4186-8992-a3a63d2d2b22" containerName="rabbitmq" containerID="cri-o://d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97" gracePeriod=604800 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.803813 4862 generic.go:334] "Generic (PLEG): container finished" podID="23891b0e-3fc7-4d2f-9abf-96b57b09ca01" containerID="36eb5af2c95a5dd49963c18b778f72ccbf03d09076fbe15508a1df0b8532d372" exitCode=137 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.826333 4862 generic.go:334] "Generic (PLEG): container finished" podID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerID="5d77e5ca098b36fcc98d53cfa5e572df42e3b1667b2235d6c42ff31fe0c66c1f" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.826417 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"989edff9-2d95-4839-9926-f5ee7dc65d19","Type":"ContainerDied","Data":"5d77e5ca098b36fcc98d53cfa5e572df42e3b1667b2235d6c42ff31fe0c66c1f"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.836342 4862 generic.go:334] "Generic (PLEG): container finished" podID="106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" containerID="d351e93457fe5a70f380ffce1f54e4bea632ff6f5419983422bc2fc2947f982d" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.836423 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" event={"ID":"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5","Type":"ContainerDied","Data":"d351e93457fe5a70f380ffce1f54e4bea632ff6f5419983422bc2fc2947f982d"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.842772 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-flbld_53d62b74-0406-47fe-8a9a-36cc6f288f27/openstack-network-exporter/0.log" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.842810 4862 generic.go:334] "Generic (PLEG): container finished" podID="53d62b74-0406-47fe-8a9a-36cc6f288f27" containerID="a1e1c2cca57dec29a126aac011c043be5314d2cdd1005a224a53029e52b33e16" exitCode=2 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.855343 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="010a34d3-73a3-4f8e-9bc3-15bb24d1d41f" path="/var/lib/kubelet/pods/010a34d3-73a3-4f8e-9bc3-15bb24d1d41f/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.856418 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0520c1cb-04b9-4790-93a1-25acd6925b94" path="/var/lib/kubelet/pods/0520c1cb-04b9-4790-93a1-25acd6925b94/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.856967 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06da04f0-f10f-4347-834a-88655ffab6fe" path="/var/lib/kubelet/pods/06da04f0-f10f-4347-834a-88655ffab6fe/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.857507 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="085bce11-bc08-44cf-a4db-d90046c891a9" path="/var/lib/kubelet/pods/085bce11-bc08-44cf-a4db-d90046c891a9/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.859718 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cfe27ee-17c6-481a-81c0-6b48211db18c" path="/var/lib/kubelet/pods/0cfe27ee-17c6-481a-81c0-6b48211db18c/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.860410 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29592660-6aa1-41b7-a96d-ac6d3db4d7c1" path="/var/lib/kubelet/pods/29592660-6aa1-41b7-a96d-ac6d3db4d7c1/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.860922 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30442d53-db04-4396-b26a-0605fa91c58f" path="/var/lib/kubelet/pods/30442d53-db04-4396-b26a-0605fa91c58f/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.862523 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ac70dca-3c3a-43fe-81ef-2677c476e9a4" path="/var/lib/kubelet/pods/3ac70dca-3c3a-43fe-81ef-2677c476e9a4/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.863009 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4be2cd5e-1345-4212-9563-449a665090f1" path="/var/lib/kubelet/pods/4be2cd5e-1345-4212-9563-449a665090f1/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.863570 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ea93c1-317d-4d0f-92a6-93872ef50d12" path="/var/lib/kubelet/pods/82ea93c1-317d-4d0f-92a6-93872ef50d12/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.865172 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b376a4e9-5456-4e8f-9c91-b222f38c30db" path="/var/lib/kubelet/pods/b376a4e9-5456-4e8f-9c91-b222f38c30db/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.865819 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eced1563-10aa-48f6-b79a-c017a3d4bba0" path="/var/lib/kubelet/pods/eced1563-10aa-48f6-b79a-c017a3d4bba0/volumes" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.866404 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-flbld" event={"ID":"53d62b74-0406-47fe-8a9a-36cc6f288f27","Type":"ContainerDied","Data":"a1e1c2cca57dec29a126aac011c043be5314d2cdd1005a224a53029e52b33e16"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871856 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871883 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871891 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871899 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871907 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871914 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871921 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871927 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871933 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871940 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.871981 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.872006 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.872015 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.872023 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.872031 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.872039 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.872048 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.872056 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.872064 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.872073 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.885046 4862 generic.go:334] "Generic (PLEG): container finished" podID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerID="a303ac97d092846a039471dffa93b8705481508ae03493b999176d9423f8a8c4" exitCode=143 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.885116 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"310ed7ff-0788-436d-b811-a4aafc7f9f5f","Type":"ContainerDied","Data":"a303ac97d092846a039471dffa93b8705481508ae03493b999176d9423f8a8c4"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.888459 4862 generic.go:334] "Generic (PLEG): container finished" podID="ca62671f-2a31-4980-93d9-2239defbfa57" containerID="ec72be9e8e9e4302a004ec5624c9e7d3c1b7137a5f5cc1b75511dcf0b235953b" exitCode=143 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.888495 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69477d9ff4-9mk7g" event={"ID":"ca62671f-2a31-4980-93d9-2239defbfa57","Type":"ContainerDied","Data":"ec72be9e8e9e4302a004ec5624c9e7d3c1b7137a5f5cc1b75511dcf0b235953b"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.890283 4862 generic.go:334] "Generic (PLEG): container finished" podID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerID="23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3" exitCode=2 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.890302 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5e37bdc4-788e-4889-9de7-682e77c3c5eb","Type":"ContainerDied","Data":"23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:20.953273 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" containerID="cri-o://c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" gracePeriod=29 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.035340 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="27c60127-ee39-4d28-89f7-e5225ddc63d7" containerName="galera" containerID="cri-o://36a419bb69dcce653bb57cc1d9e016f4dc4b3dc163be68616273503f06ee3dc5" gracePeriod=30 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.054817 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.170218 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.171776 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-flbld_53d62b74-0406-47fe-8a9a-36cc6f288f27/openstack-network-exporter/0.log" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.171855 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.173345 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_69897166-b5c9-404e-ac62-0b55f0960278/ovsdbserver-sb/0.log" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.173401 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.275973 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-config\") pod \"69897166-b5c9-404e-ac62-0b55f0960278\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276010 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovn-rundir\") pod \"53d62b74-0406-47fe-8a9a-36cc6f288f27\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276035 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-scripts\") pod \"69897166-b5c9-404e-ac62-0b55f0960278\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276059 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-combined-ca-bundle\") pod \"69897166-b5c9-404e-ac62-0b55f0960278\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276122 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovs-rundir\") pod \"53d62b74-0406-47fe-8a9a-36cc6f288f27\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276144 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-metrics-certs-tls-certs\") pod \"53d62b74-0406-47fe-8a9a-36cc6f288f27\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276171 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-metrics-certs-tls-certs\") pod \"69897166-b5c9-404e-ac62-0b55f0960278\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276155 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "53d62b74-0406-47fe-8a9a-36cc6f288f27" (UID: "53d62b74-0406-47fe-8a9a-36cc6f288f27"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276211 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-combined-ca-bundle\") pod \"53d62b74-0406-47fe-8a9a-36cc6f288f27\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276253 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"69897166-b5c9-404e-ac62-0b55f0960278\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276250 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "53d62b74-0406-47fe-8a9a-36cc6f288f27" (UID: "53d62b74-0406-47fe-8a9a-36cc6f288f27"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276302 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56qcf\" (UniqueName: \"kubernetes.io/projected/53d62b74-0406-47fe-8a9a-36cc6f288f27-kube-api-access-56qcf\") pod \"53d62b74-0406-47fe-8a9a-36cc6f288f27\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276345 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-ovsdbserver-sb-tls-certs\") pod \"69897166-b5c9-404e-ac62-0b55f0960278\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276376 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69897166-b5c9-404e-ac62-0b55f0960278-ovsdb-rundir\") pod \"69897166-b5c9-404e-ac62-0b55f0960278\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276400 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg8xp\" (UniqueName: \"kubernetes.io/projected/69897166-b5c9-404e-ac62-0b55f0960278-kube-api-access-mg8xp\") pod \"69897166-b5c9-404e-ac62-0b55f0960278\" (UID: \"69897166-b5c9-404e-ac62-0b55f0960278\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276445 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53d62b74-0406-47fe-8a9a-36cc6f288f27-config\") pod \"53d62b74-0406-47fe-8a9a-36cc6f288f27\" (UID: \"53d62b74-0406-47fe-8a9a-36cc6f288f27\") " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276691 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-config" (OuterVolumeSpecName: "config") pod "69897166-b5c9-404e-ac62-0b55f0960278" (UID: "69897166-b5c9-404e-ac62-0b55f0960278"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276799 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-scripts" (OuterVolumeSpecName: "scripts") pod "69897166-b5c9-404e-ac62-0b55f0960278" (UID: "69897166-b5c9-404e-ac62-0b55f0960278"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276874 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276887 4862 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276897 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69897166-b5c9-404e-ac62-0b55f0960278-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.276905 4862 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/53d62b74-0406-47fe-8a9a-36cc6f288f27-ovs-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.277338 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69897166-b5c9-404e-ac62-0b55f0960278-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "69897166-b5c9-404e-ac62-0b55f0960278" (UID: "69897166-b5c9-404e-ac62-0b55f0960278"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.279168 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d62b74-0406-47fe-8a9a-36cc6f288f27-config" (OuterVolumeSpecName: "config") pod "53d62b74-0406-47fe-8a9a-36cc6f288f27" (UID: "53d62b74-0406-47fe-8a9a-36cc6f288f27"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.282722 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d62b74-0406-47fe-8a9a-36cc6f288f27-kube-api-access-56qcf" (OuterVolumeSpecName: "kube-api-access-56qcf") pod "53d62b74-0406-47fe-8a9a-36cc6f288f27" (UID: "53d62b74-0406-47fe-8a9a-36cc6f288f27"). InnerVolumeSpecName "kube-api-access-56qcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.285716 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "69897166-b5c9-404e-ac62-0b55f0960278" (UID: "69897166-b5c9-404e-ac62-0b55f0960278"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.297236 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69897166-b5c9-404e-ac62-0b55f0960278-kube-api-access-mg8xp" (OuterVolumeSpecName: "kube-api-access-mg8xp") pod "69897166-b5c9-404e-ac62-0b55f0960278" (UID: "69897166-b5c9-404e-ac62-0b55f0960278"). InnerVolumeSpecName "kube-api-access-mg8xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.330119 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69897166-b5c9-404e-ac62-0b55f0960278" (UID: "69897166-b5c9-404e-ac62-0b55f0960278"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.342898 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53d62b74-0406-47fe-8a9a-36cc6f288f27" (UID: "53d62b74-0406-47fe-8a9a-36cc6f288f27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.372948 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "69897166-b5c9-404e-ac62-0b55f0960278" (UID: "69897166-b5c9-404e-ac62-0b55f0960278"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.385429 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.385470 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.385480 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56qcf\" (UniqueName: \"kubernetes.io/projected/53d62b74-0406-47fe-8a9a-36cc6f288f27-kube-api-access-56qcf\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.385490 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.385499 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69897166-b5c9-404e-ac62-0b55f0960278-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.385507 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg8xp\" (UniqueName: \"kubernetes.io/projected/69897166-b5c9-404e-ac62-0b55f0960278-kube-api-access-mg8xp\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.385515 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53d62b74-0406-47fe-8a9a-36cc6f288f27-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.385524 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.398500 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "69897166-b5c9-404e-ac62-0b55f0960278" (UID: "69897166-b5c9-404e-ac62-0b55f0960278"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.410179 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.435810 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "53d62b74-0406-47fe-8a9a-36cc6f288f27" (UID: "53d62b74-0406-47fe-8a9a-36cc6f288f27"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.439205 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="527b2bce-1186-4186-8992-a3a63d2d2b22" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.487346 4862 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53d62b74-0406-47fe-8a9a-36cc6f288f27-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.487370 4862 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69897166-b5c9-404e-ac62-0b55f0960278-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.487379 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.734529 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placemente4ca-account-delete-hcmct"] Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.741249 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican8d34-account-delete-bdd2b"] Nov 28 11:47:21 crc kubenswrapper[4862]: W1128 11:47:21.756540 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode949f4c6_f086_47e6_9a9d_b78f8b7837e7.slice/crio-ed433759386bb37f17998fc5b9f31d69a5100ecd7e4043fecffbab811e4af214 WatchSource:0}: Error finding container ed433759386bb37f17998fc5b9f31d69a5100ecd7e4043fecffbab811e4af214: Status 404 returned error can't find the container with id ed433759386bb37f17998fc5b9f31d69a5100ecd7e4043fecffbab811e4af214 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.908431 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-flbld_53d62b74-0406-47fe-8a9a-36cc6f288f27/openstack-network-exporter/0.log" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.908830 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-flbld" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.908845 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-flbld" event={"ID":"53d62b74-0406-47fe-8a9a-36cc6f288f27","Type":"ContainerDied","Data":"29b56b3656a530dddc3c588840c5692f6ddc8b025edb666c85995871c468e387"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.908952 4862 scope.go:117] "RemoveContainer" containerID="a1e1c2cca57dec29a126aac011c043be5314d2cdd1005a224a53029e52b33e16" Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.930468 4862 generic.go:334] "Generic (PLEG): container finished" podID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.930540 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hdvv6" event={"ID":"ba7cc6b9-1443-45de-84f4-4ec3b92e699c","Type":"ContainerDied","Data":"49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.939389 4862 generic.go:334] "Generic (PLEG): container finished" podID="d5203b21-897c-4331-a878-1d5c949a3e75" containerID="cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d" exitCode=143 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.939444 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6df5c754df-6lvmr" event={"ID":"d5203b21-897c-4331-a878-1d5c949a3e75","Type":"ContainerDied","Data":"cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.940916 4862 generic.go:334] "Generic (PLEG): container finished" podID="75f196ee-9a0c-414f-a629-5b270c412c24" containerID="2b26e63131e439c21df4fae7d3c514f79108333715e2829d1a0fb8c5a22b7e06" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.940953 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0c5d7-account-delete-tw6mt" event={"ID":"75f196ee-9a0c-414f-a629-5b270c412c24","Type":"ContainerDied","Data":"2b26e63131e439c21df4fae7d3c514f79108333715e2829d1a0fb8c5a22b7e06"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.944431 4862 generic.go:334] "Generic (PLEG): container finished" podID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerID="cb861c743ce00570c223afefb40f3cc5ed244b9682a5c2e0bbf669dba92ca34f" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.944463 4862 generic.go:334] "Generic (PLEG): container finished" podID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerID="b66e34c72b94210fa8e1b94676bc9d86fdb52829035fa8a84ba5bbf0ebf630c5" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.944504 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-689b7bfc65-h98mj" event={"ID":"edf3f6fa-2e74-489d-a32e-9faa5c0991ac","Type":"ContainerDied","Data":"cb861c743ce00570c223afefb40f3cc5ed244b9682a5c2e0bbf669dba92ca34f"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.944532 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-689b7bfc65-h98mj" event={"ID":"edf3f6fa-2e74-489d-a32e-9faa5c0991ac","Type":"ContainerDied","Data":"b66e34c72b94210fa8e1b94676bc9d86fdb52829035fa8a84ba5bbf0ebf630c5"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.951047 4862 generic.go:334] "Generic (PLEG): container finished" podID="73c57ace-65f0-40ab-996a-e5595851177a" containerID="e4fbe66671af80c3245a6da3f4c0fcf1f2a7319780d1ab11a9026ea78828037b" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.951115 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"73c57ace-65f0-40ab-996a-e5595851177a","Type":"ContainerDied","Data":"e4fbe66671af80c3245a6da3f4c0fcf1f2a7319780d1ab11a9026ea78828037b"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.965658 4862 generic.go:334] "Generic (PLEG): container finished" podID="77de7245-a8de-4912-b90f-669c57d8d563" containerID="adb196b1aa7dfec6e01189bb208acacebcbed93cfaed7fde33e15993b31a31f9" exitCode=143 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.965763 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77de7245-a8de-4912-b90f-669c57d8d563","Type":"ContainerDied","Data":"adb196b1aa7dfec6e01189bb208acacebcbed93cfaed7fde33e15993b31a31f9"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.967351 4862 generic.go:334] "Generic (PLEG): container finished" podID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerID="76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.967391 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc889566f-77hs4" event={"ID":"a42b2a07-23e7-4281-b69b-e47635bcccf6","Type":"ContainerDied","Data":"76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.992686 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi144c-account-delete-ls9c4"] Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.996012 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.996048 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.996079 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.996118 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e" exitCode=0 Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.996171 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.996230 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.996243 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235"} Nov 28 11:47:21 crc kubenswrapper[4862]: I1128 11:47:21.996276 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e"} Nov 28 11:47:21 crc kubenswrapper[4862]: E1128 11:47:21.998608 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 11:47:21 crc kubenswrapper[4862]: E1128 11:47:21.998670 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data podName:88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:25.998652942 +0000 UTC m=+1523.531166853 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data") pod "rabbitmq-server-0" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356") : configmap "rabbitmq-config-data" not found Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.009396 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_69897166-b5c9-404e-ac62-0b55f0960278/ovsdbserver-sb/0.log" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.009606 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.011320 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69897166-b5c9-404e-ac62-0b55f0960278","Type":"ContainerDied","Data":"a3cfa1403df510c3ae69baccf10926138e6a55c6506dc9d3b5be5956bb888c94"} Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.020646 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placemente4ca-account-delete-hcmct" event={"ID":"e949f4c6-f086-47e6-9a9d-b78f8b7837e7","Type":"ContainerStarted","Data":"ed433759386bb37f17998fc5b9f31d69a5100ecd7e4043fecffbab811e4af214"} Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.032309 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" event={"ID":"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5","Type":"ContainerDied","Data":"c83ad775ae3fa88812e0b5fcde9332a702ac484f58ea3c05ae2742a8b43e2e66"} Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.032343 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c83ad775ae3fa88812e0b5fcde9332a702ac484f58ea3c05ae2742a8b43e2e66" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.054420 4862 generic.go:334] "Generic (PLEG): container finished" podID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerID="7e91caadb693c64c084738db969219970babbd935caf36c5d56c0b2e9a460f4a" exitCode=143 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.054491 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a","Type":"ContainerDied","Data":"7e91caadb693c64c084738db969219970babbd935caf36c5d56c0b2e9a460f4a"} Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.114196 4862 generic.go:334] "Generic (PLEG): container finished" podID="27c60127-ee39-4d28-89f7-e5225ddc63d7" containerID="36a419bb69dcce653bb57cc1d9e016f4dc4b3dc163be68616273503f06ee3dc5" exitCode=0 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.114272 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"27c60127-ee39-4d28-89f7-e5225ddc63d7","Type":"ContainerDied","Data":"36a419bb69dcce653bb57cc1d9e016f4dc4b3dc163be68616273503f06ee3dc5"} Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.116320 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8d34-account-delete-bdd2b" event={"ID":"ed3d880d-6762-4d57-881d-bc585318d500","Type":"ContainerStarted","Data":"b831312cf6684723dbfa97036d2f9334dda17fd663f427687ccd6143f9604a65"} Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.118326 4862 generic.go:334] "Generic (PLEG): container finished" podID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerID="39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f" exitCode=143 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.118388 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-857784bb5b-4pc8q" event={"ID":"91afecb8-a07b-43e5-92e4-4e86cce7bcca","Type":"ContainerDied","Data":"39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f"} Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.119476 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fb8355290c76f1b9e840956935cb4162e34d4c10e8ededf078868115fa3d98d" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.121852 4862 generic.go:334] "Generic (PLEG): container finished" podID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerID="d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5" exitCode=143 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.121884 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" event={"ID":"e4f23c53-e6f8-4aca-a1e4-969a58b70a38","Type":"ContainerDied","Data":"d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5"} Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.131459 4862 scope.go:117] "RemoveContainer" containerID="670f94a7e5c8ef75b6abff70cce614bc7b826f6dcb5baf65b275f5432887d508" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.152571 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.185071 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-flbld"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.195823 4862 scope.go:117] "RemoveContainer" containerID="a1a8cae69793fcdeda0115c040ea95de4e540df38f817fe5d2600245f84cba4b" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.211240 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.218766 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-flbld"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.265764 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.267364 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b/ovsdbserver-nb/0.log" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.267446 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.273511 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.298630 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.313840 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-config\") pod \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.313877 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-swift-storage-0\") pod \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.313952 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-nb\") pod \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.313968 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw9x6\" (UniqueName: \"kubernetes.io/projected/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-kube-api-access-kw9x6\") pod \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.314013 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-svc\") pod \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.314031 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config-secret\") pod \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.314074 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-sb\") pod \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\" (UID: \"106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.314135 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config\") pod \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.314228 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxh5z\" (UniqueName: \"kubernetes.io/projected/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-kube-api-access-pxh5z\") pod \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.314275 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-combined-ca-bundle\") pod \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\" (UID: \"23891b0e-3fc7-4d2f-9abf-96b57b09ca01\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.319809 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-kube-api-access-kw9x6" (OuterVolumeSpecName: "kube-api-access-kw9x6") pod "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" (UID: "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5"). InnerVolumeSpecName "kube-api-access-kw9x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.329604 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-kube-api-access-pxh5z" (OuterVolumeSpecName: "kube-api-access-pxh5z") pod "23891b0e-3fc7-4d2f-9abf-96b57b09ca01" (UID: "23891b0e-3fc7-4d2f-9abf-96b57b09ca01"). InnerVolumeSpecName "kube-api-access-pxh5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.390624 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "23891b0e-3fc7-4d2f-9abf-96b57b09ca01" (UID: "23891b0e-3fc7-4d2f-9abf-96b57b09ca01"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.399665 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.417894 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtxql\" (UniqueName: \"kubernetes.io/projected/73c57ace-65f0-40ab-996a-e5595851177a-kube-api-access-xtxql\") pod \"73c57ace-65f0-40ab-996a-e5595851177a\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.417933 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-combined-ca-bundle\") pod \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.417976 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-config\") pod \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.418022 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdbserver-nb-tls-certs\") pod \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.418048 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-combined-ca-bundle\") pod \"73c57ace-65f0-40ab-996a-e5595851177a\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.418075 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-vencrypt-tls-certs\") pod \"73c57ace-65f0-40ab-996a-e5595851177a\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.418153 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cr4v\" (UniqueName: \"kubernetes.io/projected/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-kube-api-access-8cr4v\") pod \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.418186 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdb-rundir\") pod \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.418214 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-config-data\") pod \"73c57ace-65f0-40ab-996a-e5595851177a\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.418240 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-nova-novncproxy-tls-certs\") pod \"73c57ace-65f0-40ab-996a-e5595851177a\" (UID: \"73c57ace-65f0-40ab-996a-e5595851177a\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.418276 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-scripts\") pod \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.419408 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-metrics-certs-tls-certs\") pod \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.426569 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\" (UID: \"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.427325 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw9x6\" (UniqueName: \"kubernetes.io/projected/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-kube-api-access-kw9x6\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.427346 4862 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.427358 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxh5z\" (UniqueName: \"kubernetes.io/projected/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-kube-api-access-pxh5z\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: E1128 11:47:22.427424 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:22 crc kubenswrapper[4862]: E1128 11:47:22.427568 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data podName:527b2bce-1186-4186-8992-a3a63d2d2b22 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:26.427527916 +0000 UTC m=+1523.960041837 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data") pod "rabbitmq-cell1-server-0" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22") : configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.444002 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" (UID: "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.444918 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-scripts" (OuterVolumeSpecName: "scripts") pod "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" (UID: "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.453562 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-config" (OuterVolumeSpecName: "config") pod "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" (UID: "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.454167 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" (UID: "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.460607 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73c57ace-65f0-40ab-996a-e5595851177a-kube-api-access-xtxql" (OuterVolumeSpecName: "kube-api-access-xtxql") pod "73c57ace-65f0-40ab-996a-e5595851177a" (UID: "73c57ace-65f0-40ab-996a-e5595851177a"). InnerVolumeSpecName "kube-api-access-xtxql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.465595 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" (UID: "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.488317 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-kube-api-access-8cr4v" (OuterVolumeSpecName: "kube-api-access-8cr4v") pod "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" (UID: "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b"). InnerVolumeSpecName "kube-api-access-8cr4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.495950 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" (UID: "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.499229 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.531429 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-kolla-config\") pod \"27c60127-ee39-4d28-89f7-e5225ddc63d7\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.532867 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-default\") pod \"27c60127-ee39-4d28-89f7-e5225ddc63d7\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.534013 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-generated\") pod \"27c60127-ee39-4d28-89f7-e5225ddc63d7\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.534045 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-galera-tls-certs\") pod \"27c60127-ee39-4d28-89f7-e5225ddc63d7\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.535263 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-operator-scripts\") pod \"27c60127-ee39-4d28-89f7-e5225ddc63d7\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.535306 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"27c60127-ee39-4d28-89f7-e5225ddc63d7\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.535347 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jsjq\" (UniqueName: \"kubernetes.io/projected/27c60127-ee39-4d28-89f7-e5225ddc63d7-kube-api-access-7jsjq\") pod \"27c60127-ee39-4d28-89f7-e5225ddc63d7\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.535431 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-combined-ca-bundle\") pod \"27c60127-ee39-4d28-89f7-e5225ddc63d7\" (UID: \"27c60127-ee39-4d28-89f7-e5225ddc63d7\") " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.536577 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtxql\" (UniqueName: \"kubernetes.io/projected/73c57ace-65f0-40ab-996a-e5595851177a-kube-api-access-xtxql\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.536597 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.536626 4862 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.536634 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cr4v\" (UniqueName: \"kubernetes.io/projected/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-kube-api-access-8cr4v\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.536642 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.536651 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.536660 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.536696 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.537139 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="ceilometer-central-agent" containerID="cri-o://11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945" gracePeriod=30 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.537592 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="proxy-httpd" containerID="cri-o://35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde" gracePeriod=30 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.537642 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="sg-core" containerID="cri-o://eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee" gracePeriod=30 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.537678 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="ceilometer-notification-agent" containerID="cri-o://f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe" gracePeriod=30 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.543466 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "27c60127-ee39-4d28-89f7-e5225ddc63d7" (UID: "27c60127-ee39-4d28-89f7-e5225ddc63d7"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.543825 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "27c60127-ee39-4d28-89f7-e5225ddc63d7" (UID: "27c60127-ee39-4d28-89f7-e5225ddc63d7"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.548325 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "27c60127-ee39-4d28-89f7-e5225ddc63d7" (UID: "27c60127-ee39-4d28-89f7-e5225ddc63d7"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.550256 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "27c60127-ee39-4d28-89f7-e5225ddc63d7" (UID: "27c60127-ee39-4d28-89f7-e5225ddc63d7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.578322 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-config-data" (OuterVolumeSpecName: "config-data") pod "73c57ace-65f0-40ab-996a-e5595851177a" (UID: "73c57ace-65f0-40ab-996a-e5595851177a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.578368 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27c60127-ee39-4d28-89f7-e5225ddc63d7-kube-api-access-7jsjq" (OuterVolumeSpecName: "kube-api-access-7jsjq") pod "27c60127-ee39-4d28-89f7-e5225ddc63d7" (UID: "27c60127-ee39-4d28-89f7-e5225ddc63d7"). InnerVolumeSpecName "kube-api-access-7jsjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.579002 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23891b0e-3fc7-4d2f-9abf-96b57b09ca01" (UID: "23891b0e-3fc7-4d2f-9abf-96b57b09ca01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.591724 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.593646 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="164085a5-ab64-4c9b-b144-238e35ab3219" containerName="kube-state-metrics" containerID="cri-o://16fc0902091456b3028a91ab947a0a28c767f3d15650008ec782b9684a580f92" gracePeriod=30 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.638485 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.638521 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/27c60127-ee39-4d28-89f7-e5225ddc63d7-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.638533 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.638543 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.638553 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jsjq\" (UniqueName: \"kubernetes.io/projected/27c60127-ee39-4d28-89f7-e5225ddc63d7-kube-api-access-7jsjq\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.638561 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.638570 4862 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27c60127-ee39-4d28-89f7-e5225ddc63d7-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.645232 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronc4cb-account-delete-9jrbn"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.670076 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.670764 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="4559a88d-6d07-44d0-a829-c35fd773fcfb" containerName="memcached" containerID="cri-o://9039d4893d7c154aacf5e3d7aa83396bed1a0000865ee2be9f2d19e6c72246c2" gracePeriod=30 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.703011 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "27c60127-ee39-4d28-89f7-e5225ddc63d7" (UID: "27c60127-ee39-4d28-89f7-e5225ddc63d7"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.726722 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-gmd9b"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.744588 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7655db979f-pdkkm"] Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.744626 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.744852 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-7655db979f-pdkkm" podUID="0c30075d-2cd8-4f5a-9168-33c55459f636" containerName="keystone-api" containerID="cri-o://5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b" gracePeriod=30 Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.768346 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "23891b0e-3fc7-4d2f-9abf-96b57b09ca01" (UID: "23891b0e-3fc7-4d2f-9abf-96b57b09ca01"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.819423 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" (UID: "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.828810 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.855535 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" (UID: "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.878943 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" (UID: "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.891062 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.897598 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.897637 4862 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.897651 4862 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23891b0e-3fc7-4d2f-9abf-96b57b09ca01-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.897667 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.912885 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.162:8776/healthcheck\": read tcp 10.217.0.2:44428->10.217.0.162:8776: read: connection reset by peer" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.936326 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.965351 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23891b0e-3fc7-4d2f-9abf-96b57b09ca01" path="/var/lib/kubelet/pods/23891b0e-3fc7-4d2f-9abf-96b57b09ca01/volumes" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.965950 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d62b74-0406-47fe-8a9a-36cc6f288f27" path="/var/lib/kubelet/pods/53d62b74-0406-47fe-8a9a-36cc6f288f27/volumes" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.970726 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69897166-b5c9-404e-ac62-0b55f0960278" path="/var/lib/kubelet/pods/69897166-b5c9-404e-ac62-0b55f0960278/volumes" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.983498 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27c60127-ee39-4d28-89f7-e5225ddc63d7" (UID: "27c60127-ee39-4d28-89f7-e5225ddc63d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:22 crc kubenswrapper[4862]: I1128 11:47:22.983607 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "73c57ace-65f0-40ab-996a-e5595851177a" (UID: "73c57ace-65f0-40ab-996a-e5595851177a"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:22.999937 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:22.999965 4862 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:22.999975 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.025567 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-config" (OuterVolumeSpecName: "config") pod "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" (UID: "106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.046894 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "27c60127-ee39-4d28-89f7-e5225ddc63d7" (UID: "27c60127-ee39-4d28-89f7-e5225ddc63d7"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.050486 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73c57ace-65f0-40ab-996a-e5595851177a" (UID: "73c57ace-65f0-40ab-996a-e5595851177a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.086544 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" (UID: "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.095795 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "73c57ace-65f0-40ab-996a-e5595851177a" (UID: "73c57ace-65f0-40ab-996a-e5595851177a"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.108378 4862 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.108414 4862 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/27c60127-ee39-4d28-89f7-e5225ddc63d7-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.108423 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.108434 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.108443 4862 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c57ace-65f0-40ab-996a-e5595851177a-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.159551 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" (UID: "7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.167227 4862 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi144c-account-delete-ls9c4" secret="" err="secret \"galera-openstack-dockercfg-jvlkf\" not found" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.200627 4862 generic.go:334] "Generic (PLEG): container finished" podID="ffc2928a-7858-4117-8786-723965b126f4" containerID="2732ee7a075e6578c8913e90a93812b8f69feb263ac51f3983a02e4f5cb85ed0" exitCode=0 Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.211493 4862 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.212549 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.212616 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts podName:409c7014-982c-4c60-b374-0ff46c398db4 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:23.712599409 +0000 UTC m=+1521.245113330 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts") pod "novaapi144c-account-delete-ls9c4" (UID: "409c7014-982c-4c60-b374-0ff46c398db4") : configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.230638 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.237057 4862 generic.go:334] "Generic (PLEG): container finished" podID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerID="35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde" exitCode=0 Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.237098 4862 generic.go:334] "Generic (PLEG): container finished" podID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerID="eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee" exitCode=2 Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.248495 4862 generic.go:334] "Generic (PLEG): container finished" podID="164085a5-ab64-4c9b-b144-238e35ab3219" containerID="16fc0902091456b3028a91ab947a0a28c767f3d15650008ec782b9684a580f92" exitCode=2 Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.252257 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b/ovsdbserver-nb/0.log" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.252389 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.267341 4862 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican8d34-account-delete-bdd2b" secret="" err="secret \"galera-openstack-dockercfg-jvlkf\" not found" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.273792 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapi144c-account-delete-ls9c4" podStartSLOduration=5.273775696 podStartE2EDuration="5.273775696s" podCreationTimestamp="2025-11-28 11:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:47:23.180912755 +0000 UTC m=+1520.713426696" watchObservedRunningTime="2025-11-28 11:47:23.273775696 +0000 UTC m=+1520.806289607" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.274755 4862 generic.go:334] "Generic (PLEG): container finished" podID="e949f4c6-f086-47e6-9a9d-b78f8b7837e7" containerID="37b2c7afc1b4ec62391df0e41265269aef551c36d04b2081170ece771b12ffde" exitCode=0 Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.293199 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican8d34-account-delete-bdd2b" podStartSLOduration=5.29317936 podStartE2EDuration="5.29317936s" podCreationTimestamp="2025-11-28 11:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:47:23.279835819 +0000 UTC m=+1520.812349740" watchObservedRunningTime="2025-11-28 11:47:23.29317936 +0000 UTC m=+1520.825693281" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.298219 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b1b592a_d848_4857_894c_8f05a736dc6d.slice/crio-conmon-11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945.scope\": RecentStats: unable to find data in memory cache]" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.312815 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.312864 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts podName:ed3d880d-6762-4d57-881d-bc585318d500 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:23.812849921 +0000 UTC m=+1521.345363842 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts") pod "barbican8d34-account-delete-bdd2b" (UID: "ed3d880d-6762-4d57-881d-bc585318d500") : configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.313262 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.318830 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc646c8f9-jmhdw" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.318955 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.381256 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0682b22558ac57d011862a3e8a94d6b4334e16c0f890d30fb6ae86e9faaac01e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.396410 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0682b22558ac57d011862a3e8a94d6b4334e16c0f890d30fb6ae86e9faaac01e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.399516 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0682b22558ac57d011862a3e8a94d6b4334e16c0f890d30fb6ae86e9faaac01e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.399596 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a824cc2c-2812-4d91-9e9f-9736a27fe794" containerName="nova-scheduler-scheduler" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.406146 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi144c-account-delete-ls9c4" event={"ID":"409c7014-982c-4c60-b374-0ff46c398db4","Type":"ContainerStarted","Data":"731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.406531 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-527fq"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.406749 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-gmd9b"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.406820 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystonebc3a-account-delete-l2ccb"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.407227 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69897166-b5c9-404e-ac62-0b55f0960278" containerName="openstack-network-exporter" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.407288 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="69897166-b5c9-404e-ac62-0b55f0960278" containerName="openstack-network-exporter" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.407346 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c60127-ee39-4d28-89f7-e5225ddc63d7" containerName="galera" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.407392 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c60127-ee39-4d28-89f7-e5225ddc63d7" containerName="galera" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.407455 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c60127-ee39-4d28-89f7-e5225ddc63d7" containerName="mysql-bootstrap" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.407501 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c60127-ee39-4d28-89f7-e5225ddc63d7" containerName="mysql-bootstrap" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.407550 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73c57ace-65f0-40ab-996a-e5595851177a" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.407593 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="73c57ace-65f0-40ab-996a-e5595851177a" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.407657 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerName="openstack-network-exporter" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.407718 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerName="openstack-network-exporter" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.407786 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" containerName="dnsmasq-dns" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.407849 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" containerName="dnsmasq-dns" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.407916 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d62b74-0406-47fe-8a9a-36cc6f288f27" containerName="openstack-network-exporter" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.407971 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d62b74-0406-47fe-8a9a-36cc6f288f27" containerName="openstack-network-exporter" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.408036 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69897166-b5c9-404e-ac62-0b55f0960278" containerName="ovsdbserver-sb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.408105 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="69897166-b5c9-404e-ac62-0b55f0960278" containerName="ovsdbserver-sb" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.408158 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" containerName="init" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.408212 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" containerName="init" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.408262 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerName="ovsdbserver-nb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.408306 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerName="ovsdbserver-nb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.408515 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerName="openstack-network-exporter" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.408578 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="27c60127-ee39-4d28-89f7-e5225ddc63d7" containerName="galera" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.408627 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="69897166-b5c9-404e-ac62-0b55f0960278" containerName="ovsdbserver-sb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.410238 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d62b74-0406-47fe-8a9a-36cc6f288f27" containerName="openstack-network-exporter" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.410309 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" containerName="ovsdbserver-nb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.410359 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" containerName="dnsmasq-dns" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.410413 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="73c57ace-65f0-40ab-996a-e5595851177a" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.410460 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="69897166-b5c9-404e-ac62-0b55f0960278" containerName="openstack-network-exporter" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411032 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi144c-account-delete-ls9c4" event={"ID":"409c7014-982c-4c60-b374-0ff46c398db4","Type":"ContainerStarted","Data":"13721b6da26a5c16054dc852bfdf1a7f3824a0349e3c435a25f2b19f24b7e29f"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411123 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ffc2928a-7858-4117-8786-723965b126f4","Type":"ContainerDied","Data":"2732ee7a075e6578c8913e90a93812b8f69feb263ac51f3983a02e4f5cb85ed0"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411183 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"27c60127-ee39-4d28-89f7-e5225ddc63d7","Type":"ContainerDied","Data":"df7b749c6694f2a235cbb438e4169480bec39c6cf7788018bd36c3c069c46d73"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411237 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerDied","Data":"35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411292 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-527fq"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411354 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411442 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerDied","Data":"eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411519 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-689b7bfc65-h98mj" event={"ID":"edf3f6fa-2e74-489d-a32e-9faa5c0991ac","Type":"ContainerDied","Data":"c9ab625d86195b4566d1e54b72e2ef06cfb59eabd688fffe663d285b8371c8fc"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411576 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9ab625d86195b4566d1e54b72e2ef06cfb59eabd688fffe663d285b8371c8fc" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411624 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystonebc3a-account-delete-l2ccb"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411674 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-62jdk"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411723 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronc4cb-account-delete-9jrbn" event={"ID":"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f","Type":"ContainerStarted","Data":"ef6743975efa11f2e33cf5ac1a219f311371830c64fb99a93e655444f7507f5c"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411772 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"164085a5-ab64-4c9b-b144-238e35ab3219","Type":"ContainerDied","Data":"16fc0902091456b3028a91ab947a0a28c767f3d15650008ec782b9684a580f92"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.419543 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystonebc3a-account-delete-l2ccb"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.419641 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bc3a-account-create-update-x9czv"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.419700 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b","Type":"ContainerDied","Data":"d9f88fdac33b734215e25545d9ffc70700d4da39e0489ff0ff9f805add4a1836"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.419794 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-62jdk"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.419860 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8d34-account-delete-bdd2b" event={"ID":"ed3d880d-6762-4d57-881d-bc585318d500","Type":"ContainerStarted","Data":"3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.419922 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder68e8-account-delete-rv4w4" event={"ID":"880d930f-ea3a-4a7b-a437-5b164fbea09e","Type":"ContainerStarted","Data":"3c0d559d544a271543807e2e3c76be510d522b2484c9c8682f300b0002ab68bf"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.419978 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bc3a-account-create-update-x9czv"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.420039 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder68e8-account-delete-rv4w4"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.420129 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance056c-account-delete-cn6wm"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.420185 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placemente4ca-account-delete-hcmct" event={"ID":"e949f4c6-f086-47e6-9a9d-b78f8b7837e7","Type":"ContainerDied","Data":"37b2c7afc1b4ec62391df0e41265269aef551c36d04b2081170ece771b12ffde"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.420255 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance056c-account-delete-cn6wm" event={"ID":"71c0d0c9-ed19-41a0-a3d1-92088179bf04","Type":"ContainerStarted","Data":"ecde36974a2cbfb5129ab83ac86c44e5da8699f995d1ff184cf6a50991b0ea4e"} Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.420307 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"73c57ace-65f0-40ab-996a-e5595851177a","Type":"ContainerDied","Data":"f47d98555cb4c9a93ed28f77805a70993ecebef110883d750502c0b7280310a3"} Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.411957 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e487aac69ef596761ce6caee885f81d9e334146dc7a0a9ba8b24df61e5dc1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411637 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.411652 4862 scope.go:117] "RemoveContainer" containerID="36a419bb69dcce653bb57cc1d9e016f4dc4b3dc163be68616273503f06ee3dc5" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.419437 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.424117 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-tprgv operator-scripts], unattached volumes=[], failed to process volumes=[kube-api-access-tprgv operator-scripts]: context canceled" pod="openstack/keystonebc3a-account-delete-l2ccb" podUID="01959589-3889-4216-a92b-4e429b505a47" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.436784 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e487aac69ef596761ce6caee885f81d9e334146dc7a0a9ba8b24df61e5dc1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.447841 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e487aac69ef596761ce6caee885f81d9e334146dc7a0a9ba8b24df61e5dc1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.447882 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="ae02e889-f98f-4678-b15f-d91f711083f9" containerName="nova-cell1-conductor-conductor" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.513975 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-internal-tls-certs\") pod \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.514061 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-public-tls-certs\") pod \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.514137 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-run-httpd\") pod \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.514159 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-combined-ca-bundle\") pod \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.514221 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-log-httpd\") pod \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.514346 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-config-data\") pod \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.514501 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svqnq\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-kube-api-access-svqnq\") pod \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.514538 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-etc-swift\") pod \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\" (UID: \"edf3f6fa-2e74-489d-a32e-9faa5c0991ac\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.514852 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tprgv\" (UniqueName: \"kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv\") pod \"keystonebc3a-account-delete-l2ccb\" (UID: \"01959589-3889-4216-a92b-4e429b505a47\") " pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.514954 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts\") pod \"keystonebc3a-account-delete-l2ccb\" (UID: \"01959589-3889-4216-a92b-4e429b505a47\") " pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.516126 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "edf3f6fa-2e74-489d-a32e-9faa5c0991ac" (UID: "edf3f6fa-2e74-489d-a32e-9faa5c0991ac"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.521675 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "edf3f6fa-2e74-489d-a32e-9faa5c0991ac" (UID: "edf3f6fa-2e74-489d-a32e-9faa5c0991ac"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.525402 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "edf3f6fa-2e74-489d-a32e-9faa5c0991ac" (UID: "edf3f6fa-2e74-489d-a32e-9faa5c0991ac"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.555894 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-kube-api-access-svqnq" (OuterVolumeSpecName: "kube-api-access-svqnq") pod "edf3f6fa-2e74-489d-a32e-9faa5c0991ac" (UID: "edf3f6fa-2e74-489d-a32e-9faa5c0991ac"). InnerVolumeSpecName "kube-api-access-svqnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.577154 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-nlmd8"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.590970 4862 scope.go:117] "RemoveContainer" containerID="fcd0cc7dfd92fb5cac96b835cfe5b465117571bc62e3995b22db01397ac53877" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.591407 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-nlmd8"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.608819 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8d34-account-delete-bdd2b"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.619980 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts\") pod \"keystonebc3a-account-delete-l2ccb\" (UID: \"01959589-3889-4216-a92b-4e429b505a47\") " pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.620220 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tprgv\" (UniqueName: \"kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv\") pod \"keystonebc3a-account-delete-l2ccb\" (UID: \"01959589-3889-4216-a92b-4e429b505a47\") " pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.620642 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8d34-account-create-update-4jdj5"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.630755 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8d34-account-create-update-4jdj5"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.635639 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svqnq\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-kube-api-access-svqnq\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.635720 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.635791 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts podName:01959589-3889-4216-a92b-4e429b505a47 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:24.135773161 +0000 UTC m=+1521.668287082 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts") pod "keystonebc3a-account-delete-l2ccb" (UID: "01959589-3889-4216-a92b-4e429b505a47") : configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.636044 4862 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.636335 4862 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.636348 4862 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.646340 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-phtm9"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.647546 4862 projected.go:194] Error preparing data for projected volume kube-api-access-tprgv for pod openstack/keystonebc3a-account-delete-l2ccb: failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.647631 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv podName:01959589-3889-4216-a92b-4e429b505a47 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:24.147610829 +0000 UTC m=+1521.680124740 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-tprgv" (UniqueName: "kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv") pod "keystonebc3a-account-delete-l2ccb" (UID: "01959589-3889-4216-a92b-4e429b505a47") : failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.655758 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-phtm9"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.713186 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.714692 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.716621 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-c5d7-account-create-update-8bflf"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.722397 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.723841 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.723916 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="ovn-northd" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.736276 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0c5d7-account-delete-tw6mt"] Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.739110 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.739287 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts podName:409c7014-982c-4c60-b374-0ff46c398db4 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:24.739250735 +0000 UTC m=+1522.271764656 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts") pod "novaapi144c-account-delete-ls9c4" (UID: "409c7014-982c-4c60-b374-0ff46c398db4") : configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.751517 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-config-data" (OuterVolumeSpecName: "config-data") pod "edf3f6fa-2e74-489d-a32e-9faa5c0991ac" (UID: "edf3f6fa-2e74-489d-a32e-9faa5c0991ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.762183 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-c5d7-account-create-update-8bflf"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.786723 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "edf3f6fa-2e74-489d-a32e-9faa5c0991ac" (UID: "edf3f6fa-2e74-489d-a32e-9faa5c0991ac"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.817263 4862 scope.go:117] "RemoveContainer" containerID="e91e1d7ad1e2a8876f0520009135a4b0b1915dd5112109b61fca65064cefcd34" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.825219 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:34602->10.217.0.202:8775: read: connection reset by peer" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.825328 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:34590->10.217.0.202:8775: read: connection reset by peer" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.839859 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-certs\") pod \"164085a5-ab64-4c9b-b144-238e35ab3219\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.839947 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-config\") pod \"164085a5-ab64-4c9b-b144-238e35ab3219\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.840027 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6hmh\" (UniqueName: \"kubernetes.io/projected/164085a5-ab64-4c9b-b144-238e35ab3219-kube-api-access-q6hmh\") pod \"164085a5-ab64-4c9b-b144-238e35ab3219\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.840207 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-combined-ca-bundle\") pod \"164085a5-ab64-4c9b-b144-238e35ab3219\" (UID: \"164085a5-ab64-4c9b-b144-238e35ab3219\") " Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.840701 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.840718 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.840829 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: E1128 11:47:23.840872 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts podName:ed3d880d-6762-4d57-881d-bc585318d500 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:24.840859506 +0000 UTC m=+1522.373373427 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts") pod "barbican8d34-account-delete-bdd2b" (UID: "ed3d880d-6762-4d57-881d-bc585318d500") : configmap "openstack-scripts" not found Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.853717 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/164085a5-ab64-4c9b-b144-238e35ab3219-kube-api-access-q6hmh" (OuterVolumeSpecName: "kube-api-access-q6hmh") pod "164085a5-ab64-4c9b-b144-238e35ab3219" (UID: "164085a5-ab64-4c9b-b144-238e35ab3219"). InnerVolumeSpecName "kube-api-access-q6hmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.872693 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="f01191bd-12ff-43ea-9788-381a997a3c08" containerName="galera" containerID="cri-o://94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980" gracePeriod=30 Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.962815 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6hmh\" (UniqueName: \"kubernetes.io/projected/164085a5-ab64-4c9b-b144-238e35ab3219-kube-api-access-q6hmh\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.978133 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5jd4s"] Nov 28 11:47:23 crc kubenswrapper[4862]: I1128 11:47:23.993376 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "164085a5-ab64-4c9b-b144-238e35ab3219" (UID: "164085a5-ab64-4c9b-b144-238e35ab3219"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.041428 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5jd4s"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.054594 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi144c-account-delete-ls9c4"] Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.060488 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="821e6798c6600a5f98070c95203ea984a4b500e28b5b89da8f188ebceafa93c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.064203 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-144c-account-create-update-s9p9x"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.065507 4862 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.066309 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "edf3f6fa-2e74-489d-a32e-9faa5c0991ac" (UID: "edf3f6fa-2e74-489d-a32e-9faa5c0991ac"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.071418 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edf3f6fa-2e74-489d-a32e-9faa5c0991ac" (UID: "edf3f6fa-2e74-489d-a32e-9faa5c0991ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.072524 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-144c-account-create-update-s9p9x"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.086493 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "164085a5-ab64-4c9b-b144-238e35ab3219" (UID: "164085a5-ab64-4c9b-b144-238e35ab3219"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.086578 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-q4jzk"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.086654 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "164085a5-ab64-4c9b-b144-238e35ab3219" (UID: "164085a5-ab64-4c9b-b144-238e35ab3219"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.090140 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-q4jzk"] Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.095311 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="821e6798c6600a5f98070c95203ea984a4b500e28b5b89da8f188ebceafa93c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.109137 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-e4ca-account-create-update-hsh77"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.121150 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-e4ca-account-create-update-hsh77"] Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.121300 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="821e6798c6600a5f98070c95203ea984a4b500e28b5b89da8f188ebceafa93c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.121353 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="01f7b6ff-ee76-4961-9baf-930743da6143" containerName="nova-cell0-conductor-conductor" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.121315 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placemente4ca-account-delete-hcmct"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.170850 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tprgv\" (UniqueName: \"kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv\") pod \"keystonebc3a-account-delete-l2ccb\" (UID: \"01959589-3889-4216-a92b-4e429b505a47\") " pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.170982 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts\") pod \"keystonebc3a-account-delete-l2ccb\" (UID: \"01959589-3889-4216-a92b-4e429b505a47\") " pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.171064 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.171075 4862 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/164085a5-ab64-4c9b-b144-238e35ab3219-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.171086 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.171114 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3f6fa-2e74-489d-a32e-9faa5c0991ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.171200 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.171262 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts podName:01959589-3889-4216-a92b-4e429b505a47 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:25.171233937 +0000 UTC m=+1522.703747848 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts") pod "keystonebc3a-account-delete-l2ccb" (UID: "01959589-3889-4216-a92b-4e429b505a47") : configmap "openstack-scripts" not found Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.173835 4862 projected.go:194] Error preparing data for projected volume kube-api-access-tprgv for pod openstack/keystonebc3a-account-delete-l2ccb: failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.173887 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv podName:01959589-3889-4216-a92b-4e429b505a47 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:25.173859452 +0000 UTC m=+1522.706373373 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-tprgv" (UniqueName: "kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv") pod "keystonebc3a-account-delete-l2ccb" (UID: "01959589-3889-4216-a92b-4e429b505a47") : failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.249161 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-857784bb5b-4pc8q" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:33460->10.217.0.157:9311: read: connection reset by peer" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.249420 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-857784bb5b-4pc8q" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:33466->10.217.0.157:9311: read: connection reset by peer" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.259152 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-7v76b"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.297705 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-7v76b"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.324336 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronc4cb-account-delete-9jrbn"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.363705 4862 generic.go:334] "Generic (PLEG): container finished" podID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerID="11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945" exitCode=0 Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.363805 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerDied","Data":"11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.365156 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c4cb-account-create-update-fmhl6"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.381418 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ffc2928a-7858-4117-8786-723965b126f4","Type":"ContainerDied","Data":"a1504d4191ee33f529cd3c01feba8563b3fed4553cd518928314cae338b4ede1"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.381460 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1504d4191ee33f529cd3c01feba8563b3fed4553cd518928314cae338b4ede1" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.383447 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c4cb-account-create-update-fmhl6"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.399346 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronc4cb-account-delete-9jrbn" event={"ID":"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f","Type":"ContainerStarted","Data":"9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.399820 4862 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutronc4cb-account-delete-9jrbn" secret="" err="secret \"galera-openstack-dockercfg-jvlkf\" not found" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.404579 4862 generic.go:334] "Generic (PLEG): container finished" podID="ca62671f-2a31-4980-93d9-2239defbfa57" containerID="a7c34e84cffc71a4ccf07325940e732e6d5ffc06d972e6e93c18df3db794e0b9" exitCode=0 Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.404644 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69477d9ff4-9mk7g" event={"ID":"ca62671f-2a31-4980-93d9-2239defbfa57","Type":"ContainerDied","Data":"a7c34e84cffc71a4ccf07325940e732e6d5ffc06d972e6e93c18df3db794e0b9"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.408195 4862 generic.go:334] "Generic (PLEG): container finished" podID="01f7b6ff-ee76-4961-9baf-930743da6143" containerID="821e6798c6600a5f98070c95203ea984a4b500e28b5b89da8f188ebceafa93c7" exitCode=0 Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.408236 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"01f7b6ff-ee76-4961-9baf-930743da6143","Type":"ContainerDied","Data":"821e6798c6600a5f98070c95203ea984a4b500e28b5b89da8f188ebceafa93c7"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.422188 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutronc4cb-account-delete-9jrbn" podStartSLOduration=5.422172621 podStartE2EDuration="5.422172621s" podCreationTimestamp="2025-11-28 11:47:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:47:24.418901858 +0000 UTC m=+1521.951415779" watchObservedRunningTime="2025-11-28 11:47:24.422172621 +0000 UTC m=+1521.954686542" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.424967 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.426628 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"164085a5-ab64-4c9b-b144-238e35ab3219","Type":"ContainerDied","Data":"d4d25375805a2c1a55dbdedaa7447bc1b0e831a818fc3da9ce359c0fb27a9eab"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.473147 4862 generic.go:334] "Generic (PLEG): container finished" podID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerID="b622cbdd5687d59d52eeed34f730456b8b029c86f0df4d63c0373c42dc4992e7" exitCode=0 Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.473228 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a","Type":"ContainerDied","Data":"b622cbdd5687d59d52eeed34f730456b8b029c86f0df4d63c0373c42dc4992e7"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.477458 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"310ed7ff-0788-436d-b811-a4aafc7f9f5f","Type":"ContainerDied","Data":"15217408eaf0c5ecfdbf05adb57c81819bf35870f42f21b0ecf1bd8c4c5ff8a8"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.477414 4862 generic.go:334] "Generic (PLEG): container finished" podID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerID="15217408eaf0c5ecfdbf05adb57c81819bf35870f42f21b0ecf1bd8c4c5ff8a8" exitCode=0 Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.485674 4862 generic.go:334] "Generic (PLEG): container finished" podID="4559a88d-6d07-44d0-a829-c35fd773fcfb" containerID="9039d4893d7c154aacf5e3d7aa83396bed1a0000865ee2be9f2d19e6c72246c2" exitCode=0 Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.485750 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4559a88d-6d07-44d0-a829-c35fd773fcfb","Type":"ContainerDied","Data":"9039d4893d7c154aacf5e3d7aa83396bed1a0000865ee2be9f2d19e6c72246c2"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.485773 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4559a88d-6d07-44d0-a829-c35fd773fcfb","Type":"ContainerDied","Data":"9ee7bf639c6fcd4781e3e20eed08767a0f373457bcd9bab0bb0a4c93cd017667"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.485784 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ee7bf639c6fcd4781e3e20eed08767a0f373457bcd9bab0bb0a4c93cd017667" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.492117 4862 generic.go:334] "Generic (PLEG): container finished" podID="77de7245-a8de-4912-b90f-669c57d8d563" containerID="69388439d162f0fc46186cc6f4708492782333d0c2f8c07c9238c3381aaf99ad" exitCode=0 Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.492168 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77de7245-a8de-4912-b90f-669c57d8d563","Type":"ContainerDied","Data":"69388439d162f0fc46186cc6f4708492782333d0c2f8c07c9238c3381aaf99ad"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.501509 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0c5d7-account-delete-tw6mt" event={"ID":"75f196ee-9a0c-414f-a629-5b270c412c24","Type":"ContainerDied","Data":"ac599f52e26cdf7e02a4c590fcf1a44aa33c68d4a7f0ba3b99172dad869beaf0"} Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.501587 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac599f52e26cdf7e02a4c590fcf1a44aa33c68d4a7f0ba3b99172dad869beaf0" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.501747 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-689b7bfc65-h98mj" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.502137 4862 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican8d34-account-delete-bdd2b" secret="" err="secret \"galera-openstack-dockercfg-jvlkf\" not found" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.502850 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.503079 4862 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi144c-account-delete-ls9c4" secret="" err="secret \"galera-openstack-dockercfg-jvlkf\" not found" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.539294 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-5pnwp"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.549338 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-5pnwp"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.565517 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-056c-account-create-update-4t96j"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.576860 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-056c-account-create-update-4t96j"] Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.581820 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.581885 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts podName:dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:47:25.08187028 +0000 UTC m=+1522.614384191 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts") pod "neutronc4cb-account-delete-9jrbn" (UID: "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f") : configmap "openstack-scripts" not found Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.587290 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance056c-account-delete-cn6wm"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.620711 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.634706 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.651186 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.659322 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.672558 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.684876 4862 scope.go:117] "RemoveContainer" containerID="ee8d7e9a14861e17f472a1f44420408da6cefecb358b29bd4e8a5166746780c2" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.707391 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.748832 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.749274 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.752826 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-8ndh9"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.762125 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-8ndh9"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.784340 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc646c8f9-jmhdw"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.791323 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data-custom\") pod \"ffc2928a-7858-4117-8786-723965b126f4\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.791400 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-scripts\") pod \"ffc2928a-7858-4117-8786-723965b126f4\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.791448 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27g6d\" (UniqueName: \"kubernetes.io/projected/ffc2928a-7858-4117-8786-723965b126f4-kube-api-access-27g6d\") pod \"ffc2928a-7858-4117-8786-723965b126f4\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.791491 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-internal-tls-certs\") pod \"ffc2928a-7858-4117-8786-723965b126f4\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.791545 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc2928a-7858-4117-8786-723965b126f4-logs\") pod \"ffc2928a-7858-4117-8786-723965b126f4\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.791663 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffc2928a-7858-4117-8786-723965b126f4-etc-machine-id\") pod \"ffc2928a-7858-4117-8786-723965b126f4\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.791737 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-public-tls-certs\") pod \"ffc2928a-7858-4117-8786-723965b126f4\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.791765 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data\") pod \"ffc2928a-7858-4117-8786-723965b126f4\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.791976 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-combined-ca-bundle\") pod \"ffc2928a-7858-4117-8786-723965b126f4\" (UID: \"ffc2928a-7858-4117-8786-723965b126f4\") " Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.792427 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.793073 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts podName:409c7014-982c-4c60-b374-0ff46c398db4 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:26.793056539 +0000 UTC m=+1524.325570460 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts") pod "novaapi144c-account-delete-ls9c4" (UID: "409c7014-982c-4c60-b374-0ff46c398db4") : configmap "openstack-scripts" not found Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.794734 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffc2928a-7858-4117-8786-723965b126f4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ffc2928a-7858-4117-8786-723965b126f4" (UID: "ffc2928a-7858-4117-8786-723965b126f4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.795239 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc2928a-7858-4117-8786-723965b126f4-logs" (OuterVolumeSpecName: "logs") pod "ffc2928a-7858-4117-8786-723965b126f4" (UID: "ffc2928a-7858-4117-8786-723965b126f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.807193 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-scripts" (OuterVolumeSpecName: "scripts") pod "ffc2928a-7858-4117-8786-723965b126f4" (UID: "ffc2928a-7858-4117-8786-723965b126f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.818582 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ffc2928a-7858-4117-8786-723965b126f4" (UID: "ffc2928a-7858-4117-8786-723965b126f4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.838054 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc646c8f9-jmhdw"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.902603 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e88fe61-1181-4385-8d0a-d99483e8112f" path="/var/lib/kubelet/pods/0e88fe61-1181-4385-8d0a-d99483e8112f/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.903931 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc2928a-7858-4117-8786-723965b126f4-kube-api-access-27g6d" (OuterVolumeSpecName: "kube-api-access-27g6d") pod "ffc2928a-7858-4117-8786-723965b126f4" (UID: "ffc2928a-7858-4117-8786-723965b126f4"). InnerVolumeSpecName "kube-api-access-27g6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.904014 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5" path="/var/lib/kubelet/pods/106e1c4f-afb5-4f94-9a9e-e3bb02e17eb5/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.905764 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-config-data\") pod \"4559a88d-6d07-44d0-a829-c35fd773fcfb\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.905876 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-combined-ca-bundle\") pod \"4559a88d-6d07-44d0-a829-c35fd773fcfb\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.905917 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75f196ee-9a0c-414f-a629-5b270c412c24-operator-scripts\") pod \"75f196ee-9a0c-414f-a629-5b270c412c24\" (UID: \"75f196ee-9a0c-414f-a629-5b270c412c24\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.906032 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcjnt\" (UniqueName: \"kubernetes.io/projected/4559a88d-6d07-44d0-a829-c35fd773fcfb-kube-api-access-hcjnt\") pod \"4559a88d-6d07-44d0-a829-c35fd773fcfb\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.906066 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-kolla-config\") pod \"4559a88d-6d07-44d0-a829-c35fd773fcfb\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.906179 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-memcached-tls-certs\") pod \"4559a88d-6d07-44d0-a829-c35fd773fcfb\" (UID: \"4559a88d-6d07-44d0-a829-c35fd773fcfb\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.906280 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzsw9\" (UniqueName: \"kubernetes.io/projected/75f196ee-9a0c-414f-a629-5b270c412c24-kube-api-access-lzsw9\") pod \"75f196ee-9a0c-414f-a629-5b270c412c24\" (UID: \"75f196ee-9a0c-414f-a629-5b270c412c24\") " Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.908943 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75f196ee-9a0c-414f-a629-5b270c412c24-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "75f196ee-9a0c-414f-a629-5b270c412c24" (UID: "75f196ee-9a0c-414f-a629-5b270c412c24"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.909798 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-config-data" (OuterVolumeSpecName: "config-data") pod "4559a88d-6d07-44d0-a829-c35fd773fcfb" (UID: "4559a88d-6d07-44d0-a829-c35fd773fcfb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.910346 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "4559a88d-6d07-44d0-a829-c35fd773fcfb" (UID: "4559a88d-6d07-44d0-a829-c35fd773fcfb"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.912896 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de88799-efb7-4f0e-8c25-ace191d5885e" path="/var/lib/kubelet/pods/1de88799-efb7-4f0e-8c25-ace191d5885e/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.918341 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.921375 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.921397 4862 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffc2928a-7858-4117-8786-723965b126f4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.921409 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75f196ee-9a0c-414f-a629-5b270c412c24-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.921419 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.921431 4862 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4559a88d-6d07-44d0-a829-c35fd773fcfb-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.921440 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.921449 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27g6d\" (UniqueName: \"kubernetes.io/projected/ffc2928a-7858-4117-8786-723965b126f4-kube-api-access-27g6d\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.921458 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc2928a-7858-4117-8786-723965b126f4-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.921594 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.921641 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts podName:ed3d880d-6762-4d57-881d-bc585318d500 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:26.9216212 +0000 UTC m=+1524.454135121 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts") pod "barbican8d34-account-delete-bdd2b" (UID: "ed3d880d-6762-4d57-881d-bc585318d500") : configmap "openstack-scripts" not found Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.923820 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.926237 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.926268 4862 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.933525 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.941946 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ec7a6dc-a28c-4239-b112-6139ae089cee" path="/var/lib/kubelet/pods/1ec7a6dc-a28c-4239-b112-6139ae089cee/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.943980 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27c60127-ee39-4d28-89f7-e5225ddc63d7" path="/var/lib/kubelet/pods/27c60127-ee39-4d28-89f7-e5225ddc63d7/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.944943 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.945675 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3772cd33-8cc8-47a2-8453-f2395f011239" path="/var/lib/kubelet/pods/3772cd33-8cc8-47a2-8453-f2395f011239/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.946422 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aa74c33-9ed9-4d71-925e-e5140feebbff" path="/var/lib/kubelet/pods/3aa74c33-9ed9-4d71-925e-e5140feebbff/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.948225 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a1ff8a7-99a3-4655-b491-0a10f01c2ee4" path="/var/lib/kubelet/pods/4a1ff8a7-99a3-4655-b491-0a10f01c2ee4/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.948975 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57dc2822-b0b6-45d4-92de-5f601ebcf48a" path="/var/lib/kubelet/pods/57dc2822-b0b6-45d4-92de-5f601ebcf48a/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.950369 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4559a88d-6d07-44d0-a829-c35fd773fcfb-kube-api-access-hcjnt" (OuterVolumeSpecName: "kube-api-access-hcjnt") pod "4559a88d-6d07-44d0-a829-c35fd773fcfb" (UID: "4559a88d-6d07-44d0-a829-c35fd773fcfb"). InnerVolumeSpecName "kube-api-access-hcjnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.950514 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:24 crc kubenswrapper[4862]: E1128 11:47:24.950580 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.956950 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5987550f-5082-4c09-a428-8ddc77c38598" path="/var/lib/kubelet/pods/5987550f-5082-4c09-a428-8ddc77c38598/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.957710 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="705faed3-031b-4ec7-bfd7-b5ac626c98b3" path="/var/lib/kubelet/pods/705faed3-031b-4ec7-bfd7-b5ac626c98b3/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.959623 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b" path="/var/lib/kubelet/pods/7ad1b0b7-0bf2-4f48-8787-7bc6477d9f1b/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.960697 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f196ee-9a0c-414f-a629-5b270c412c24-kube-api-access-lzsw9" (OuterVolumeSpecName: "kube-api-access-lzsw9") pod "75f196ee-9a0c-414f-a629-5b270c412c24" (UID: "75f196ee-9a0c-414f-a629-5b270c412c24"). InnerVolumeSpecName "kube-api-access-lzsw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.962800 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88411d96-8709-4e1a-9042-a84a33008a18" path="/var/lib/kubelet/pods/88411d96-8709-4e1a-9042-a84a33008a18/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.963756 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8abae65f-f5e4-4b16-a8f0-8892145fd6bf" path="/var/lib/kubelet/pods/8abae65f-f5e4-4b16-a8f0-8892145fd6bf/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.964489 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1e1bd20-ac01-449f-b10c-aeba36373db9" path="/var/lib/kubelet/pods/a1e1bd20-ac01-449f-b10c-aeba36373db9/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.965762 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a51e2e8d-a77a-4017-90bd-1e2dfdc8f527" path="/var/lib/kubelet/pods/a51e2e8d-a77a-4017-90bd-1e2dfdc8f527/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.966763 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adfa85c5-0924-4b39-bc3d-d2d884b19161" path="/var/lib/kubelet/pods/adfa85c5-0924-4b39-bc3d-d2d884b19161/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.969912 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d990f484-bffa-4bfc-8c45-b00c5514512a" path="/var/lib/kubelet/pods/d990f484-bffa-4bfc-8c45-b00c5514512a/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.970809 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f558e8a3-94d7-4bb8-a390-71d7ed87bb9d" path="/var/lib/kubelet/pods/f558e8a3-94d7-4bb8-a390-71d7ed87bb9d/volumes" Nov 28 11:47:24 crc kubenswrapper[4862]: I1128 11:47:24.973280 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff113c73-a3c6-4300-93f2-fed5c3a1040f" path="/var/lib/kubelet/pods/ff113c73-a3c6-4300-93f2-fed5c3a1040f/volumes" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.026307 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcjnt\" (UniqueName: \"kubernetes.io/projected/4559a88d-6d07-44d0-a829-c35fd773fcfb-kube-api-access-hcjnt\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.026332 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzsw9\" (UniqueName: \"kubernetes.io/projected/75f196ee-9a0c-414f-a629-5b270c412c24-kube-api-access-lzsw9\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.035247 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hkgwr" podUID="b1bdf176-776a-4211-8b43-ff90225de25a" containerName="ovn-controller" probeResult="failure" output=< Nov 28 11:47:25 crc kubenswrapper[4862]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Nov 28 11:47:25 crc kubenswrapper[4862]: > Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.063652 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffc2928a-7858-4117-8786-723965b126f4" (UID: "ffc2928a-7858-4117-8786-723965b126f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.103694 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.199:3000/\": dial tcp 10.217.0.199:3000: connect: connection refused" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.127587 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: E1128 11:47:25.127661 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:25 crc kubenswrapper[4862]: E1128 11:47:25.127708 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts podName:dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:47:26.127692463 +0000 UTC m=+1523.660206384 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts") pod "neutronc4cb-account-delete-9jrbn" (UID: "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f") : configmap "openstack-scripts" not found Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.133307 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6dc889566f-77hs4" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.150:9696/\": dial tcp 10.217.0.150:9696: connect: connection refused" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.188237 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data" (OuterVolumeSpecName: "config-data") pod "ffc2928a-7858-4117-8786-723965b126f4" (UID: "ffc2928a-7858-4117-8786-723965b126f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.200389 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ffc2928a-7858-4117-8786-723965b126f4" (UID: "ffc2928a-7858-4117-8786-723965b126f4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.203235 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "4559a88d-6d07-44d0-a829-c35fd773fcfb" (UID: "4559a88d-6d07-44d0-a829-c35fd773fcfb"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.231612 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tprgv\" (UniqueName: \"kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv\") pod \"keystonebc3a-account-delete-l2ccb\" (UID: \"01959589-3889-4216-a92b-4e429b505a47\") " pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.231716 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4559a88d-6d07-44d0-a829-c35fd773fcfb" (UID: "4559a88d-6d07-44d0-a829-c35fd773fcfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.231742 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts\") pod \"keystonebc3a-account-delete-l2ccb\" (UID: \"01959589-3889-4216-a92b-4e429b505a47\") " pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:25 crc kubenswrapper[4862]: E1128 11:47:25.231840 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:25 crc kubenswrapper[4862]: E1128 11:47:25.231891 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts podName:01959589-3889-4216-a92b-4e429b505a47 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:27.231875727 +0000 UTC m=+1524.764389648 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts") pod "keystonebc3a-account-delete-l2ccb" (UID: "01959589-3889-4216-a92b-4e429b505a47") : configmap "openstack-scripts" not found Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.232046 4862 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.232066 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4559a88d-6d07-44d0-a829-c35fd773fcfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.232078 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.232107 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: E1128 11:47:25.236920 4862 projected.go:194] Error preparing data for projected volume kube-api-access-tprgv for pod openstack/keystonebc3a-account-delete-l2ccb: failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 11:47:25 crc kubenswrapper[4862]: E1128 11:47:25.236974 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv podName:01959589-3889-4216-a92b-4e429b505a47 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:27.236961203 +0000 UTC m=+1524.769475124 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-tprgv" (UniqueName: "kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv") pod "keystonebc3a-account-delete-l2ccb" (UID: "01959589-3889-4216-a92b-4e429b505a47") : failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.260190 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ffc2928a-7858-4117-8786-723965b126f4" (UID: "ffc2928a-7858-4117-8786-723965b126f4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.284529 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder68e8-account-delete-rv4w4"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.284564 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-68e8-account-create-update-4zwm5"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.284579 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-68e8-account-create-update-4zwm5"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.284593 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.284603 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.284621 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.284630 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.284639 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-689b7bfc65-h98mj"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.284651 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-689b7bfc65-h98mj"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.293229 4862 scope.go:117] "RemoveContainer" containerID="e4fbe66671af80c3245a6da3f4c0fcf1f2a7319780d1ab11a9026ea78828037b" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.298936 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.320575 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.339931 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-config-data\") pod \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.340688 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t2jd\" (UniqueName: \"kubernetes.io/projected/77de7245-a8de-4912-b90f-669c57d8d563-kube-api-access-9t2jd\") pod \"77de7245-a8de-4912-b90f-669c57d8d563\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.340988 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.341029 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-nova-metadata-tls-certs\") pod \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.341142 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-combined-ca-bundle\") pod \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.341285 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-combined-ca-bundle\") pod \"77de7245-a8de-4912-b90f-669c57d8d563\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.341344 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b584j\" (UniqueName: \"kubernetes.io/projected/310ed7ff-0788-436d-b811-a4aafc7f9f5f-kube-api-access-b584j\") pod \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.341390 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ed7ff-0788-436d-b811-a4aafc7f9f5f-logs\") pod \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\" (UID: \"310ed7ff-0788-436d-b811-a4aafc7f9f5f\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.342470 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-internal-tls-certs\") pod \"77de7245-a8de-4912-b90f-669c57d8d563\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.342582 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77de7245-a8de-4912-b90f-669c57d8d563-logs\") pod \"77de7245-a8de-4912-b90f-669c57d8d563\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.342625 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-public-tls-certs\") pod \"77de7245-a8de-4912-b90f-669c57d8d563\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.342657 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-config-data\") pod \"77de7245-a8de-4912-b90f-669c57d8d563\" (UID: \"77de7245-a8de-4912-b90f-669c57d8d563\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.343624 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffc2928a-7858-4117-8786-723965b126f4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.343752 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/310ed7ff-0788-436d-b811-a4aafc7f9f5f-logs" (OuterVolumeSpecName: "logs") pod "310ed7ff-0788-436d-b811-a4aafc7f9f5f" (UID: "310ed7ff-0788-436d-b811-a4aafc7f9f5f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.344338 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77de7245-a8de-4912-b90f-669c57d8d563-logs" (OuterVolumeSpecName: "logs") pod "77de7245-a8de-4912-b90f-669c57d8d563" (UID: "77de7245-a8de-4912-b90f-669c57d8d563"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.356086 4862 scope.go:117] "RemoveContainer" containerID="16fc0902091456b3028a91ab947a0a28c767f3d15650008ec782b9684a580f92" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.356260 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77de7245-a8de-4912-b90f-669c57d8d563-kube-api-access-9t2jd" (OuterVolumeSpecName: "kube-api-access-9t2jd") pod "77de7245-a8de-4912-b90f-669c57d8d563" (UID: "77de7245-a8de-4912-b90f-669c57d8d563"). InnerVolumeSpecName "kube-api-access-9t2jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.369734 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/310ed7ff-0788-436d-b811-a4aafc7f9f5f-kube-api-access-b584j" (OuterVolumeSpecName: "kube-api-access-b584j") pod "310ed7ff-0788-436d-b811-a4aafc7f9f5f" (UID: "310ed7ff-0788-436d-b811-a4aafc7f9f5f"). InnerVolumeSpecName "kube-api-access-b584j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.371235 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.375795 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.376907 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.386964 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.408452 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "310ed7ff-0788-436d-b811-a4aafc7f9f5f" (UID: "310ed7ff-0788-436d-b811-a4aafc7f9f5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.419422 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77de7245-a8de-4912-b90f-669c57d8d563" (UID: "77de7245-a8de-4912-b90f-669c57d8d563"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.435996 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "310ed7ff-0788-436d-b811-a4aafc7f9f5f" (UID: "310ed7ff-0788-436d-b811-a4aafc7f9f5f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.442011 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-config-data" (OuterVolumeSpecName: "config-data") pod "310ed7ff-0788-436d-b811-a4aafc7f9f5f" (UID: "310ed7ff-0788-436d-b811-a4aafc7f9f5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.445805 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-scripts\") pod \"e8ef59ee-556c-4735-8cdf-e140aa608f86\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.445872 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"e8ef59ee-556c-4735-8cdf-e140aa608f86\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.445891 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-config-data\") pod \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.445909 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data\") pod \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.445954 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-internal-tls-certs\") pod \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.445975 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-logs\") pod \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.446029 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rtl9\" (UniqueName: \"kubernetes.io/projected/e8ef59ee-556c-4735-8cdf-e140aa608f86-kube-api-access-6rtl9\") pod \"e8ef59ee-556c-4735-8cdf-e140aa608f86\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.446056 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-combined-ca-bundle\") pod \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.446194 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljnj4\" (UniqueName: \"kubernetes.io/projected/ca62671f-2a31-4980-93d9-2239defbfa57-kube-api-access-ljnj4\") pod \"ca62671f-2a31-4980-93d9-2239defbfa57\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.446215 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-httpd-run\") pod \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.446233 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9hrc\" (UniqueName: \"kubernetes.io/projected/91afecb8-a07b-43e5-92e4-4e86cce7bcca-kube-api-access-c9hrc\") pod \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.446299 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26rfx\" (UniqueName: \"kubernetes.io/projected/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-kube-api-access-26rfx\") pod \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.446317 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-combined-ca-bundle\") pod \"ca62671f-2a31-4980-93d9-2239defbfa57\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.446367 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-scripts\") pod \"ca62671f-2a31-4980-93d9-2239defbfa57\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.446389 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-combined-ca-bundle\") pod \"e8ef59ee-556c-4735-8cdf-e140aa608f86\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447226 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-internal-tls-certs\") pod \"ca62671f-2a31-4980-93d9-2239defbfa57\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447281 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-internal-tls-certs\") pod \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447316 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-httpd-run\") pod \"e8ef59ee-556c-4735-8cdf-e140aa608f86\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447367 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-combined-ca-bundle\") pod \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447402 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62671f-2a31-4980-93d9-2239defbfa57-logs\") pod \"ca62671f-2a31-4980-93d9-2239defbfa57\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447418 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-combined-ca-bundle\") pod \"01f7b6ff-ee76-4961-9baf-930743da6143\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447440 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447455 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q5gq\" (UniqueName: \"kubernetes.io/projected/01f7b6ff-ee76-4961-9baf-930743da6143-kube-api-access-9q5gq\") pod \"01f7b6ff-ee76-4961-9baf-930743da6143\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447474 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-public-tls-certs\") pod \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447491 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-scripts\") pod \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\" (UID: \"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447529 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-config-data\") pod \"01f7b6ff-ee76-4961-9baf-930743da6143\" (UID: \"01f7b6ff-ee76-4961-9baf-930743da6143\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447554 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91afecb8-a07b-43e5-92e4-4e86cce7bcca-logs\") pod \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447568 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-public-tls-certs\") pod \"e8ef59ee-556c-4735-8cdf-e140aa608f86\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447602 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data-custom\") pod \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\" (UID: \"91afecb8-a07b-43e5-92e4-4e86cce7bcca\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447637 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-config-data\") pod \"ca62671f-2a31-4980-93d9-2239defbfa57\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447659 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-public-tls-certs\") pod \"ca62671f-2a31-4980-93d9-2239defbfa57\" (UID: \"ca62671f-2a31-4980-93d9-2239defbfa57\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447676 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-config-data\") pod \"e8ef59ee-556c-4735-8cdf-e140aa608f86\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.447704 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-logs\") pod \"e8ef59ee-556c-4735-8cdf-e140aa608f86\" (UID: \"e8ef59ee-556c-4735-8cdf-e140aa608f86\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.448367 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.448405 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.448415 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b584j\" (UniqueName: \"kubernetes.io/projected/310ed7ff-0788-436d-b811-a4aafc7f9f5f-kube-api-access-b584j\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.448426 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ed7ff-0788-436d-b811-a4aafc7f9f5f-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.448435 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77de7245-a8de-4912-b90f-669c57d8d563-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.448443 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.448476 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t2jd\" (UniqueName: \"kubernetes.io/projected/77de7245-a8de-4912-b90f-669c57d8d563-kube-api-access-9t2jd\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.448485 4862 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/310ed7ff-0788-436d-b811-a4aafc7f9f5f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.448795 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-logs" (OuterVolumeSpecName: "logs") pod "e8ef59ee-556c-4735-8cdf-e140aa608f86" (UID: "e8ef59ee-556c-4735-8cdf-e140aa608f86"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.467353 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-logs" (OuterVolumeSpecName: "logs") pod "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" (UID: "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.476632 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca62671f-2a31-4980-93d9-2239defbfa57-logs" (OuterVolumeSpecName: "logs") pod "ca62671f-2a31-4980-93d9-2239defbfa57" (UID: "ca62671f-2a31-4980-93d9-2239defbfa57"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.482646 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-kube-api-access-26rfx" (OuterVolumeSpecName: "kube-api-access-26rfx") pod "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" (UID: "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a"). InnerVolumeSpecName "kube-api-access-26rfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.496649 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91afecb8-a07b-43e5-92e4-4e86cce7bcca-logs" (OuterVolumeSpecName: "logs") pod "91afecb8-a07b-43e5-92e4-4e86cce7bcca" (UID: "91afecb8-a07b-43e5-92e4-4e86cce7bcca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.497015 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-scripts" (OuterVolumeSpecName: "scripts") pod "e8ef59ee-556c-4735-8cdf-e140aa608f86" (UID: "e8ef59ee-556c-4735-8cdf-e140aa608f86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.497084 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "77de7245-a8de-4912-b90f-669c57d8d563" (UID: "77de7245-a8de-4912-b90f-669c57d8d563"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.509065 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" (UID: "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.509180 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e8ef59ee-556c-4735-8cdf-e140aa608f86" (UID: "e8ef59ee-556c-4735-8cdf-e140aa608f86"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.509651 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-config-data" (OuterVolumeSpecName: "config-data") pod "77de7245-a8de-4912-b90f-669c57d8d563" (UID: "77de7245-a8de-4912-b90f-669c57d8d563"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.510230 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "91afecb8-a07b-43e5-92e4-4e86cce7bcca" (UID: "91afecb8-a07b-43e5-92e4-4e86cce7bcca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.517862 4862 generic.go:334] "Generic (PLEG): container finished" podID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerID="8031be805d28d00e73f2d74cac7b5638f7feeaa4fd5d9e786af7a60f92caf4cc" exitCode=0 Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.517915 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"989edff9-2d95-4839-9926-f5ee7dc65d19","Type":"ContainerDied","Data":"8031be805d28d00e73f2d74cac7b5638f7feeaa4fd5d9e786af7a60f92caf4cc"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.518012 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "e8ef59ee-556c-4735-8cdf-e140aa608f86" (UID: "e8ef59ee-556c-4735-8cdf-e140aa608f86"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.518114 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01f7b6ff-ee76-4961-9baf-930743da6143-kube-api-access-9q5gq" (OuterVolumeSpecName: "kube-api-access-9q5gq") pod "01f7b6ff-ee76-4961-9baf-930743da6143" (UID: "01f7b6ff-ee76-4961-9baf-930743da6143"). InnerVolumeSpecName "kube-api-access-9q5gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.518169 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-scripts" (OuterVolumeSpecName: "scripts") pod "ca62671f-2a31-4980-93d9-2239defbfa57" (UID: "ca62671f-2a31-4980-93d9-2239defbfa57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.518219 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91afecb8-a07b-43e5-92e4-4e86cce7bcca-kube-api-access-c9hrc" (OuterVolumeSpecName: "kube-api-access-c9hrc") pod "91afecb8-a07b-43e5-92e4-4e86cce7bcca" (UID: "91afecb8-a07b-43e5-92e4-4e86cce7bcca"). InnerVolumeSpecName "kube-api-access-c9hrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.523607 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-config-data" (OuterVolumeSpecName: "config-data") pod "01f7b6ff-ee76-4961-9baf-930743da6143" (UID: "01f7b6ff-ee76-4961-9baf-930743da6143"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.525264 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ef59ee-556c-4735-8cdf-e140aa608f86-kube-api-access-6rtl9" (OuterVolumeSpecName: "kube-api-access-6rtl9") pod "e8ef59ee-556c-4735-8cdf-e140aa608f86" (UID: "e8ef59ee-556c-4735-8cdf-e140aa608f86"). InnerVolumeSpecName "kube-api-access-6rtl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.553325 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" (UID: "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.553338 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-scripts" (OuterVolumeSpecName: "scripts") pod "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" (UID: "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554755 4862 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554779 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9hrc\" (UniqueName: \"kubernetes.io/projected/91afecb8-a07b-43e5-92e4-4e86cce7bcca-kube-api-access-c9hrc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554789 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26rfx\" (UniqueName: \"kubernetes.io/projected/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-kube-api-access-26rfx\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554797 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554805 4862 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554813 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62671f-2a31-4980-93d9-2239defbfa57-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554836 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554847 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q5gq\" (UniqueName: \"kubernetes.io/projected/01f7b6ff-ee76-4961-9baf-930743da6143-kube-api-access-9q5gq\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554857 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554866 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554877 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554886 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91afecb8-a07b-43e5-92e4-4e86cce7bcca-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554896 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554905 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554913 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8ef59ee-556c-4735-8cdf-e140aa608f86-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554921 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554936 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554947 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.554955 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rtl9\" (UniqueName: \"kubernetes.io/projected/e8ef59ee-556c-4735-8cdf-e140aa608f86-kube-api-access-6rtl9\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.555121 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.555402 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77de7245-a8de-4912-b90f-669c57d8d563","Type":"ContainerDied","Data":"2fac0b3dcfcd10093df99f352ea83a49a5118ccdde56ae84f862b2f3813021dd"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.555449 4862 scope.go:117] "RemoveContainer" containerID="69388439d162f0fc46186cc6f4708492782333d0c2f8c07c9238c3381aaf99ad" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.559109 4862 generic.go:334] "Generic (PLEG): container finished" podID="ae02e889-f98f-4678-b15f-d91f711083f9" containerID="e487aac69ef596761ce6caee885f81d9e334146dc7a0a9ba8b24df61e5dc1e8f" exitCode=0 Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.559159 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ae02e889-f98f-4678-b15f-d91f711083f9","Type":"ContainerDied","Data":"e487aac69ef596761ce6caee885f81d9e334146dc7a0a9ba8b24df61e5dc1e8f"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.560237 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance056c-account-delete-cn6wm" event={"ID":"71c0d0c9-ed19-41a0-a3d1-92088179bf04","Type":"ContainerStarted","Data":"83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.560339 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance056c-account-delete-cn6wm" podUID="71c0d0c9-ed19-41a0-a3d1-92088179bf04" containerName="mariadb-account-delete" containerID="cri-o://83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772" gracePeriod=30 Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.563549 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca62671f-2a31-4980-93d9-2239defbfa57-kube-api-access-ljnj4" (OuterVolumeSpecName: "kube-api-access-ljnj4") pod "ca62671f-2a31-4980-93d9-2239defbfa57" (UID: "ca62671f-2a31-4980-93d9-2239defbfa57"). InnerVolumeSpecName "kube-api-access-ljnj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.563563 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69477d9ff4-9mk7g" event={"ID":"ca62671f-2a31-4980-93d9-2239defbfa57","Type":"ContainerDied","Data":"d890a1ed5cddce80a6edf3b669ba45708f46953b7f4f422b9d8c1feb9ef58997"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.563609 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69477d9ff4-9mk7g" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.566330 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"01f7b6ff-ee76-4961-9baf-930743da6143","Type":"ContainerDied","Data":"8d91061053a251e8e8dd0c4b382c70337c48209fecb8ce9c1d7b99feffd01a9e"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.566396 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.572046 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"310ed7ff-0788-436d-b811-a4aafc7f9f5f","Type":"ContainerDied","Data":"b9548b98bbb6b54ca79db5174ace520be3e184d53112db741e8ed1cd25393c9a"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.572204 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.576769 4862 generic.go:334] "Generic (PLEG): container finished" podID="a824cc2c-2812-4d91-9e9f-9736a27fe794" containerID="0682b22558ac57d011862a3e8a94d6b4334e16c0f890d30fb6ae86e9faaac01e" exitCode=0 Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.576986 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a824cc2c-2812-4d91-9e9f-9736a27fe794","Type":"ContainerDied","Data":"0682b22558ac57d011862a3e8a94d6b4334e16c0f890d30fb6ae86e9faaac01e"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.582868 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder68e8-account-delete-rv4w4" event={"ID":"880d930f-ea3a-4a7b-a437-5b164fbea09e","Type":"ContainerStarted","Data":"4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.582955 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance056c-account-delete-cn6wm" podStartSLOduration=6.58293676 podStartE2EDuration="6.58293676s" podCreationTimestamp="2025-11-28 11:47:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:47:25.572035469 +0000 UTC m=+1523.104549380" watchObservedRunningTime="2025-11-28 11:47:25.58293676 +0000 UTC m=+1523.115450681" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.583227 4862 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder68e8-account-delete-rv4w4" secret="" err="secret \"galera-openstack-dockercfg-jvlkf\" not found" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.586289 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-857784bb5b-4pc8q" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.586302 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-857784bb5b-4pc8q" event={"ID":"91afecb8-a07b-43e5-92e4-4e86cce7bcca","Type":"ContainerDied","Data":"773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.586283 4862 generic.go:334] "Generic (PLEG): container finished" podID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerID="773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92" exitCode=0 Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.586455 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-857784bb5b-4pc8q" event={"ID":"91afecb8-a07b-43e5-92e4-4e86cce7bcca","Type":"ContainerDied","Data":"807c30986f85c37de51a7b74d0c8e012c3c5178a714e192cb5cce33e8d23e0f1"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.587818 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01f7b6ff-ee76-4961-9baf-930743da6143" (UID: "01f7b6ff-ee76-4961-9baf-930743da6143"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.590011 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a","Type":"ContainerDied","Data":"b332abc3cb712724bc373cf5882882de6d6b2dc3faa817a5caf4646ff373152b"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.590032 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.597447 4862 generic.go:334] "Generic (PLEG): container finished" podID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerID="ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8" exitCode=0 Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.597539 4862 scope.go:117] "RemoveContainer" containerID="adb196b1aa7dfec6e01189bb208acacebcbed93cfaed7fde33e15993b31a31f9" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.597569 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.597675 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.597699 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e8ef59ee-556c-4735-8cdf-e140aa608f86","Type":"ContainerDied","Data":"ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.597720 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e8ef59ee-556c-4735-8cdf-e140aa608f86","Type":"ContainerDied","Data":"a99d9afba970fedc6cd31e5248110433094a405b5a46447f466d6177833cc195"} Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.597760 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonebc3a-account-delete-l2ccb" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.598679 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0c5d7-account-delete-tw6mt" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.599199 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.599405 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.599643 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutronc4cb-account-delete-9jrbn" podUID="dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f" containerName="mariadb-account-delete" containerID="cri-o://9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b" gracePeriod=30 Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.599927 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapi144c-account-delete-ls9c4" podUID="409c7014-982c-4c60-b374-0ff46c398db4" containerName="mariadb-account-delete" containerID="cri-o://731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a" gracePeriod=30 Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.600003 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican8d34-account-delete-bdd2b" podUID="ed3d880d-6762-4d57-881d-bc585318d500" containerName="mariadb-account-delete" containerID="cri-o://3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698" gracePeriod=30 Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.603177 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder68e8-account-delete-rv4w4" podStartSLOduration=6.603153307 podStartE2EDuration="6.603153307s" podCreationTimestamp="2025-11-28 11:47:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 11:47:25.596552529 +0000 UTC m=+1523.129066450" watchObservedRunningTime="2025-11-28 11:47:25.603153307 +0000 UTC m=+1523.135667238" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.619751 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.630350 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.634940 4862 scope.go:117] "RemoveContainer" containerID="a7c34e84cffc71a4ccf07325940e732e6d5ffc06d972e6e93c18df3db794e0b9" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.643187 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.644152 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-config-data" (OuterVolumeSpecName: "config-data") pod "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" (UID: "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.653111 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "77de7245-a8de-4912-b90f-669c57d8d563" (UID: "77de7245-a8de-4912-b90f-669c57d8d563"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.655635 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-config-data\") pod \"a824cc2c-2812-4d91-9e9f-9736a27fe794\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.655738 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data\") pod \"989edff9-2d95-4839-9926-f5ee7dc65d19\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.655801 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pccc2\" (UniqueName: \"kubernetes.io/projected/989edff9-2d95-4839-9926-f5ee7dc65d19-kube-api-access-pccc2\") pod \"989edff9-2d95-4839-9926-f5ee7dc65d19\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.655840 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-combined-ca-bundle\") pod \"a824cc2c-2812-4d91-9e9f-9736a27fe794\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.655888 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/989edff9-2d95-4839-9926-f5ee7dc65d19-etc-machine-id\") pod \"989edff9-2d95-4839-9926-f5ee7dc65d19\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.655928 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf7gr\" (UniqueName: \"kubernetes.io/projected/a824cc2c-2812-4d91-9e9f-9736a27fe794-kube-api-access-kf7gr\") pod \"a824cc2c-2812-4d91-9e9f-9736a27fe794\" (UID: \"a824cc2c-2812-4d91-9e9f-9736a27fe794\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.655961 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data-custom\") pod \"989edff9-2d95-4839-9926-f5ee7dc65d19\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.655986 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jznv\" (UniqueName: \"kubernetes.io/projected/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-kube-api-access-9jznv\") pod \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\" (UID: \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.656167 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-combined-ca-bundle\") pod \"989edff9-2d95-4839-9926-f5ee7dc65d19\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.656190 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-scripts\") pod \"989edff9-2d95-4839-9926-f5ee7dc65d19\" (UID: \"989edff9-2d95-4839-9926-f5ee7dc65d19\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.656218 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-operator-scripts\") pod \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\" (UID: \"e949f4c6-f086-47e6-9a9d-b78f8b7837e7\") " Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.657049 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f7b6ff-ee76-4961-9baf-930743da6143-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.657074 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77de7245-a8de-4912-b90f-669c57d8d563-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.657121 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.657137 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljnj4\" (UniqueName: \"kubernetes.io/projected/ca62671f-2a31-4980-93d9-2239defbfa57-kube-api-access-ljnj4\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.661859 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/989edff9-2d95-4839-9926-f5ee7dc65d19-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "989edff9-2d95-4839-9926-f5ee7dc65d19" (UID: "989edff9-2d95-4839-9926-f5ee7dc65d19"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: E1128 11:47:25.662615 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:25 crc kubenswrapper[4862]: E1128 11:47:25.662665 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts podName:880d930f-ea3a-4a7b-a437-5b164fbea09e nodeName:}" failed. No retries permitted until 2025-11-28 11:47:26.162650916 +0000 UTC m=+1523.695164837 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts") pod "cinder68e8-account-delete-rv4w4" (UID: "880d930f-ea3a-4a7b-a437-5b164fbea09e") : configmap "openstack-scripts" not found Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.662844 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e949f4c6-f086-47e6-9a9d-b78f8b7837e7" (UID: "e949f4c6-f086-47e6-9a9d-b78f8b7837e7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.663012 4862 scope.go:117] "RemoveContainer" containerID="ec72be9e8e9e4302a004ec5624c9e7d3c1b7137a5f5cc1b75511dcf0b235953b" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.664993 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.667574 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.675336 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-scripts" (OuterVolumeSpecName: "scripts") pod "989edff9-2d95-4839-9926-f5ee7dc65d19" (UID: "989edff9-2d95-4839-9926-f5ee7dc65d19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.675361 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" (UID: "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.675391 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a824cc2c-2812-4d91-9e9f-9736a27fe794-kube-api-access-kf7gr" (OuterVolumeSpecName: "kube-api-access-kf7gr") pod "a824cc2c-2812-4d91-9e9f-9736a27fe794" (UID: "a824cc2c-2812-4d91-9e9f-9736a27fe794"). InnerVolumeSpecName "kube-api-access-kf7gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.675463 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "989edff9-2d95-4839-9926-f5ee7dc65d19" (UID: "989edff9-2d95-4839-9926-f5ee7dc65d19"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.675482 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-kube-api-access-9jznv" (OuterVolumeSpecName: "kube-api-access-9jznv") pod "e949f4c6-f086-47e6-9a9d-b78f8b7837e7" (UID: "e949f4c6-f086-47e6-9a9d-b78f8b7837e7"). InnerVolumeSpecName "kube-api-access-9jznv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.675514 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/989edff9-2d95-4839-9926-f5ee7dc65d19-kube-api-access-pccc2" (OuterVolumeSpecName: "kube-api-access-pccc2") pod "989edff9-2d95-4839-9926-f5ee7dc65d19" (UID: "989edff9-2d95-4839-9926-f5ee7dc65d19"). InnerVolumeSpecName "kube-api-access-pccc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.687884 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "91afecb8-a07b-43e5-92e4-4e86cce7bcca" (UID: "91afecb8-a07b-43e5-92e4-4e86cce7bcca"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.691318 4862 scope.go:117] "RemoveContainer" containerID="821e6798c6600a5f98070c95203ea984a4b500e28b5b89da8f188ebceafa93c7" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.726621 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.736506 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e8ef59ee-556c-4735-8cdf-e140aa608f86" (UID: "e8ef59ee-556c-4735-8cdf-e140aa608f86"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.737073 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystonebc3a-account-delete-l2ccb"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.756914 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystonebc3a-account-delete-l2ccb"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768818 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768853 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768867 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768879 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768891 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768902 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pccc2\" (UniqueName: \"kubernetes.io/projected/989edff9-2d95-4839-9926-f5ee7dc65d19-kube-api-access-pccc2\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768913 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768922 4862 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/989edff9-2d95-4839-9926-f5ee7dc65d19-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768936 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768946 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf7gr\" (UniqueName: \"kubernetes.io/projected/a824cc2c-2812-4d91-9e9f-9736a27fe794-kube-api-access-kf7gr\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768957 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.768968 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jznv\" (UniqueName: \"kubernetes.io/projected/e949f4c6-f086-47e6-9a9d-b78f8b7837e7-kube-api-access-9jznv\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.769141 4862 scope.go:117] "RemoveContainer" containerID="15217408eaf0c5ecfdbf05adb57c81819bf35870f42f21b0ecf1bd8c4c5ff8a8" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.769277 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0c5d7-account-delete-tw6mt"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.777307 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0c5d7-account-delete-tw6mt"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.779456 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca62671f-2a31-4980-93d9-2239defbfa57" (UID: "ca62671f-2a31-4980-93d9-2239defbfa57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.803183 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91afecb8-a07b-43e5-92e4-4e86cce7bcca" (UID: "91afecb8-a07b-43e5-92e4-4e86cce7bcca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.803833 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.807320 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8ef59ee-556c-4735-8cdf-e140aa608f86" (UID: "e8ef59ee-556c-4735-8cdf-e140aa608f86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.835555 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-config-data" (OuterVolumeSpecName: "config-data") pod "ca62671f-2a31-4980-93d9-2239defbfa57" (UID: "ca62671f-2a31-4980-93d9-2239defbfa57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.855448 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data" (OuterVolumeSpecName: "config-data") pod "91afecb8-a07b-43e5-92e4-4e86cce7bcca" (UID: "91afecb8-a07b-43e5-92e4-4e86cce7bcca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.882467 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tprgv\" (UniqueName: \"kubernetes.io/projected/01959589-3889-4216-a92b-4e429b505a47-kube-api-access-tprgv\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.885497 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.885546 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.885558 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.885568 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.885578 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01959589-3889-4216-a92b-4e429b505a47-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.885591 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.902464 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.909876 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "989edff9-2d95-4839-9926-f5ee7dc65d19" (UID: "989edff9-2d95-4839-9926-f5ee7dc65d19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.927183 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-config-data" (OuterVolumeSpecName: "config-data") pod "e8ef59ee-556c-4735-8cdf-e140aa608f86" (UID: "e8ef59ee-556c-4735-8cdf-e140aa608f86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.932720 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-config-data" (OuterVolumeSpecName: "config-data") pod "a824cc2c-2812-4d91-9e9f-9736a27fe794" (UID: "a824cc2c-2812-4d91-9e9f-9736a27fe794"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.952773 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a824cc2c-2812-4d91-9e9f-9736a27fe794" (UID: "a824cc2c-2812-4d91-9e9f-9736a27fe794"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:25 crc kubenswrapper[4862]: I1128 11:47:25.987231 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:25.998841 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:25.998876 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8ef59ee-556c-4735-8cdf-e140aa608f86-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:25.998886 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:25.998895 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a824cc2c-2812-4d91-9e9f-9736a27fe794-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.000351 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.000404 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data podName:88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:34.000386918 +0000 UTC m=+1531.532900849 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data") pod "rabbitmq-server-0" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356") : configmap "rabbitmq-config-data" not found Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.003254 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "91afecb8-a07b-43e5-92e4-4e86cce7bcca" (UID: "91afecb8-a07b-43e5-92e4-4e86cce7bcca"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.004757 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" (UID: "4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.023285 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.053719 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ca62671f-2a31-4980-93d9-2239defbfa57" (UID: "ca62671f-2a31-4980-93d9-2239defbfa57"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.072540 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data" (OuterVolumeSpecName: "config-data") pod "989edff9-2d95-4839-9926-f5ee7dc65d19" (UID: "989edff9-2d95-4839-9926-f5ee7dc65d19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.100649 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.100668 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.100677 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989edff9-2d95-4839-9926-f5ee7dc65d19-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.100686 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91afecb8-a07b-43e5-92e4-4e86cce7bcca-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.106289 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ca62671f-2a31-4980-93d9-2239defbfa57" (UID: "ca62671f-2a31-4980-93d9-2239defbfa57"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.204770 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca62671f-2a31-4980-93d9-2239defbfa57-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.204849 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.204896 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts podName:880d930f-ea3a-4a7b-a437-5b164fbea09e nodeName:}" failed. No retries permitted until 2025-11-28 11:47:27.204880436 +0000 UTC m=+1524.737394357 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts") pod "cinder68e8-account-delete-rv4w4" (UID: "880d930f-ea3a-4a7b-a437-5b164fbea09e") : configmap "openstack-scripts" not found Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.205113 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.205141 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts podName:dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:47:28.205133493 +0000 UTC m=+1525.737647414 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts") pod "neutronc4cb-account-delete-9jrbn" (UID: "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f") : configmap "openstack-scripts" not found Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.269014 4862 scope.go:117] "RemoveContainer" containerID="a303ac97d092846a039471dffa93b8705481508ae03493b999176d9423f8a8c4" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.275328 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.297901 4862 scope.go:117] "RemoveContainer" containerID="773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.315391 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-combined-ca-bundle\") pod \"ae02e889-f98f-4678-b15f-d91f711083f9\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.315518 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-config-data\") pod \"ae02e889-f98f-4678-b15f-d91f711083f9\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.315559 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xqpn\" (UniqueName: \"kubernetes.io/projected/ae02e889-f98f-4678-b15f-d91f711083f9-kube-api-access-8xqpn\") pod \"ae02e889-f98f-4678-b15f-d91f711083f9\" (UID: \"ae02e889-f98f-4678-b15f-d91f711083f9\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.321879 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae02e889-f98f-4678-b15f-d91f711083f9-kube-api-access-8xqpn" (OuterVolumeSpecName: "kube-api-access-8xqpn") pod "ae02e889-f98f-4678-b15f-d91f711083f9" (UID: "ae02e889-f98f-4678-b15f-d91f711083f9"). InnerVolumeSpecName "kube-api-access-8xqpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.327044 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-69477d9ff4-9mk7g"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.340273 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-69477d9ff4-9mk7g"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.349765 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-857784bb5b-4pc8q"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.368341 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-config-data" (OuterVolumeSpecName: "config-data") pod "ae02e889-f98f-4678-b15f-d91f711083f9" (UID: "ae02e889-f98f-4678-b15f-d91f711083f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.369638 4862 scope.go:117] "RemoveContainer" containerID="39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.369687 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-857784bb5b-4pc8q"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.377411 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.383924 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.393404 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.401466 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.413948 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae02e889-f98f-4678-b15f-d91f711083f9" (UID: "ae02e889-f98f-4678-b15f-d91f711083f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.414874 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.417816 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.417840 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae02e889-f98f-4678-b15f-d91f711083f9-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.417850 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xqpn\" (UniqueName: \"kubernetes.io/projected/ae02e889-f98f-4678-b15f-d91f711083f9-kube-api-access-8xqpn\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.419941 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.422440 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.426388 4862 scope.go:117] "RemoveContainer" containerID="773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92" Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.426842 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92\": container with ID starting with 773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92 not found: ID does not exist" containerID="773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.426906 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92"} err="failed to get container status \"773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92\": rpc error: code = NotFound desc = could not find container \"773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92\": container with ID starting with 773b78b0d3cf12e275bde8abecfa06df01b0a65fcce872e72369ecf9d61fdb92 not found: ID does not exist" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.426928 4862 scope.go:117] "RemoveContainer" containerID="39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f" Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.427420 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f\": container with ID starting with 39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f not found: ID does not exist" containerID="39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.427476 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f"} err="failed to get container status \"39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f\": rpc error: code = NotFound desc = could not find container \"39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f\": container with ID starting with 39e76098a71176626d5a2ec45f661fe03b5d2f662afc8b5867b6a1094badc64f not found: ID does not exist" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.427499 4862 scope.go:117] "RemoveContainer" containerID="b622cbdd5687d59d52eeed34f730456b8b029c86f0df4d63c0373c42dc4992e7" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.432480 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.439196 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.463973 4862 scope.go:117] "RemoveContainer" containerID="7e91caadb693c64c084738db969219970babbd935caf36c5d56c0b2e9a460f4a" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.484156 4862 scope.go:117] "RemoveContainer" containerID="ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.487800 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.515886 4862 scope.go:117] "RemoveContainer" containerID="236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518400 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-fernet-keys\") pod \"0c30075d-2cd8-4f5a-9168-33c55459f636\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518429 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-galera-tls-certs\") pod \"f01191bd-12ff-43ea-9788-381a997a3c08\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518467 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-public-tls-certs\") pod \"0c30075d-2cd8-4f5a-9168-33c55459f636\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518517 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-generated\") pod \"f01191bd-12ff-43ea-9788-381a997a3c08\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518543 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-credential-keys\") pod \"0c30075d-2cd8-4f5a-9168-33c55459f636\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518564 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44qrp\" (UniqueName: \"kubernetes.io/projected/f01191bd-12ff-43ea-9788-381a997a3c08-kube-api-access-44qrp\") pod \"f01191bd-12ff-43ea-9788-381a997a3c08\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518585 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-config-data\") pod \"0c30075d-2cd8-4f5a-9168-33c55459f636\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518612 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-scripts\") pod \"0c30075d-2cd8-4f5a-9168-33c55459f636\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518630 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-kolla-config\") pod \"f01191bd-12ff-43ea-9788-381a997a3c08\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518649 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b9k7\" (UniqueName: \"kubernetes.io/projected/0c30075d-2cd8-4f5a-9168-33c55459f636-kube-api-access-9b9k7\") pod \"0c30075d-2cd8-4f5a-9168-33c55459f636\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518682 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-combined-ca-bundle\") pod \"0c30075d-2cd8-4f5a-9168-33c55459f636\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518730 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-internal-tls-certs\") pod \"0c30075d-2cd8-4f5a-9168-33c55459f636\" (UID: \"0c30075d-2cd8-4f5a-9168-33c55459f636\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518772 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-default\") pod \"f01191bd-12ff-43ea-9788-381a997a3c08\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518804 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-combined-ca-bundle\") pod \"f01191bd-12ff-43ea-9788-381a997a3c08\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518833 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-operator-scripts\") pod \"f01191bd-12ff-43ea-9788-381a997a3c08\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.518868 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"f01191bd-12ff-43ea-9788-381a997a3c08\" (UID: \"f01191bd-12ff-43ea-9788-381a997a3c08\") " Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.519250 4862 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.519293 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data podName:527b2bce-1186-4186-8992-a3a63d2d2b22 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:34.519280952 +0000 UTC m=+1532.051794883 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data") pod "rabbitmq-cell1-server-0" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22") : configmap "rabbitmq-cell1-config-data" not found Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.519771 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "f01191bd-12ff-43ea-9788-381a997a3c08" (UID: "f01191bd-12ff-43ea-9788-381a997a3c08"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.519868 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "f01191bd-12ff-43ea-9788-381a997a3c08" (UID: "f01191bd-12ff-43ea-9788-381a997a3c08"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.520313 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "f01191bd-12ff-43ea-9788-381a997a3c08" (UID: "f01191bd-12ff-43ea-9788-381a997a3c08"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.523241 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f01191bd-12ff-43ea-9788-381a997a3c08" (UID: "f01191bd-12ff-43ea-9788-381a997a3c08"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.577234 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0c30075d-2cd8-4f5a-9168-33c55459f636" (UID: "0c30075d-2cd8-4f5a-9168-33c55459f636"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.578285 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c30075d-2cd8-4f5a-9168-33c55459f636-kube-api-access-9b9k7" (OuterVolumeSpecName: "kube-api-access-9b9k7") pod "0c30075d-2cd8-4f5a-9168-33c55459f636" (UID: "0c30075d-2cd8-4f5a-9168-33c55459f636"). InnerVolumeSpecName "kube-api-access-9b9k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.578398 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0c30075d-2cd8-4f5a-9168-33c55459f636" (UID: "0c30075d-2cd8-4f5a-9168-33c55459f636"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.579261 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-scripts" (OuterVolumeSpecName: "scripts") pod "0c30075d-2cd8-4f5a-9168-33c55459f636" (UID: "0c30075d-2cd8-4f5a-9168-33c55459f636"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.580231 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f01191bd-12ff-43ea-9788-381a997a3c08" (UID: "f01191bd-12ff-43ea-9788-381a997a3c08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.580240 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c30075d-2cd8-4f5a-9168-33c55459f636" (UID: "0c30075d-2cd8-4f5a-9168-33c55459f636"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.587194 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f01191bd-12ff-43ea-9788-381a997a3c08-kube-api-access-44qrp" (OuterVolumeSpecName: "kube-api-access-44qrp") pod "f01191bd-12ff-43ea-9788-381a997a3c08" (UID: "f01191bd-12ff-43ea-9788-381a997a3c08"). InnerVolumeSpecName "kube-api-access-44qrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.591972 4862 scope.go:117] "RemoveContainer" containerID="ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8" Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.596360 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8\": container with ID starting with ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8 not found: ID does not exist" containerID="ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.596416 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8"} err="failed to get container status \"ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8\": rpc error: code = NotFound desc = could not find container \"ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8\": container with ID starting with ac25beeeeefb99bf986ddcbe024e95e37a7c7b3a268103cda88ddd13c3f438a8 not found: ID does not exist" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.596443 4862 scope.go:117] "RemoveContainer" containerID="236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb" Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.597319 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb\": container with ID starting with 236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb not found: ID does not exist" containerID="236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.597359 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb"} err="failed to get container status \"236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb\": rpc error: code = NotFound desc = could not find container \"236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb\": container with ID starting with 236bb0773ebcd43ced6bc58bfb07f2b18873238f0ff7b67f445a3cac0057febb not found: ID does not exist" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.597747 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "f01191bd-12ff-43ea-9788-381a997a3c08" (UID: "f01191bd-12ff-43ea-9788-381a997a3c08"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.612793 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-config-data" (OuterVolumeSpecName: "config-data") pod "0c30075d-2cd8-4f5a-9168-33c55459f636" (UID: "0c30075d-2cd8-4f5a-9168-33c55459f636"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621023 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621047 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621056 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621076 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621098 4862 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621107 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f01191bd-12ff-43ea-9788-381a997a3c08-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621118 4862 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621126 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44qrp\" (UniqueName: \"kubernetes.io/projected/f01191bd-12ff-43ea-9788-381a997a3c08-kube-api-access-44qrp\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621135 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621143 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621150 4862 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f01191bd-12ff-43ea-9788-381a997a3c08-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621158 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b9k7\" (UniqueName: \"kubernetes.io/projected/0c30075d-2cd8-4f5a-9168-33c55459f636-kube-api-access-9b9k7\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621166 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621537 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placemente4ca-account-delete-hcmct" event={"ID":"e949f4c6-f086-47e6-9a9d-b78f8b7837e7","Type":"ContainerDied","Data":"ed433759386bb37f17998fc5b9f31d69a5100ecd7e4043fecffbab811e4af214"} Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.621571 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed433759386bb37f17998fc5b9f31d69a5100ecd7e4043fecffbab811e4af214" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.622160 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placemente4ca-account-delete-hcmct" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.645282 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"989edff9-2d95-4839-9926-f5ee7dc65d19","Type":"ContainerDied","Data":"8e6ad7cefc6df96185f6ce389da91df490b069298a8999fae01055de3567d518"} Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.645330 4862 scope.go:117] "RemoveContainer" containerID="5d77e5ca098b36fcc98d53cfa5e572df42e3b1667b2235d6c42ff31fe0c66c1f" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.645475 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.652977 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.655733 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0c30075d-2cd8-4f5a-9168-33c55459f636" (UID: "0c30075d-2cd8-4f5a-9168-33c55459f636"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.656208 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "f01191bd-12ff-43ea-9788-381a997a3c08" (UID: "f01191bd-12ff-43ea-9788-381a997a3c08"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.660033 4862 generic.go:334] "Generic (PLEG): container finished" podID="f01191bd-12ff-43ea-9788-381a997a3c08" containerID="94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980" exitCode=0 Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.660131 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f01191bd-12ff-43ea-9788-381a997a3c08","Type":"ContainerDied","Data":"94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980"} Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.660162 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f01191bd-12ff-43ea-9788-381a997a3c08","Type":"ContainerDied","Data":"4123c28387f62633cc6d3b318c5591f34816fcef70e767fa769f385b325aedc3"} Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.660235 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.665765 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0c30075d-2cd8-4f5a-9168-33c55459f636" (UID: "0c30075d-2cd8-4f5a-9168-33c55459f636"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.683423 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a824cc2c-2812-4d91-9e9f-9736a27fe794","Type":"ContainerDied","Data":"d10eaea45792fe4cd29893a74175c64063f071faa3833b0635963d49903e4e16"} Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.683594 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.692696 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ae02e889-f98f-4678-b15f-d91f711083f9","Type":"ContainerDied","Data":"e1e0952a769d2d2ea44aeded428a163cea3def05adf8bb7b2dcb71f595cd6959"} Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.692849 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.705108 4862 generic.go:334] "Generic (PLEG): container finished" podID="0c30075d-2cd8-4f5a-9168-33c55459f636" containerID="5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b" exitCode=0 Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.705175 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7655db979f-pdkkm" event={"ID":"0c30075d-2cd8-4f5a-9168-33c55459f636","Type":"ContainerDied","Data":"5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b"} Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.705199 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7655db979f-pdkkm" event={"ID":"0c30075d-2cd8-4f5a-9168-33c55459f636","Type":"ContainerDied","Data":"b508df719851845b47a039b4e7bfe0523253f825c7e85f7002bc75e3ce377fbc"} Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.705257 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7655db979f-pdkkm" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.719088 4862 scope.go:117] "RemoveContainer" containerID="8031be805d28d00e73f2d74cac7b5638f7feeaa4fd5d9e786af7a60f92caf4cc" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.720963 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placemente4ca-account-delete-hcmct"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.723953 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.723988 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c30075d-2cd8-4f5a-9168-33c55459f636-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.724000 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.724138 4862 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01191bd-12ff-43ea-9788-381a997a3c08-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.742000 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placemente4ca-account-delete-hcmct"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.748502 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.751825 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder68e8-account-delete-rv4w4" podUID="880d930f-ea3a-4a7b-a437-5b164fbea09e" containerName="mariadb-account-delete" containerID="cri-o://4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790" gracePeriod=30 Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.759060 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.764889 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.770034 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.826180 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:26 crc kubenswrapper[4862]: E1128 11:47:26.826276 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts podName:409c7014-982c-4c60-b374-0ff46c398db4 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:30.826255476 +0000 UTC m=+1528.358769397 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts") pod "novaapi144c-account-delete-ls9c4" (UID: "409c7014-982c-4c60-b374-0ff46c398db4") : configmap "openstack-scripts" not found Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.842177 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="73c57ace-65f0-40ab-996a-e5595851177a" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.194:6080/vnc_lite.html\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.858613 4862 scope.go:117] "RemoveContainer" containerID="94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.864539 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01959589-3889-4216-a92b-4e429b505a47" path="/var/lib/kubelet/pods/01959589-3889-4216-a92b-4e429b505a47/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.865017 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01f7b6ff-ee76-4961-9baf-930743da6143" path="/var/lib/kubelet/pods/01f7b6ff-ee76-4961-9baf-930743da6143/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.866378 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="164085a5-ab64-4c9b-b144-238e35ab3219" path="/var/lib/kubelet/pods/164085a5-ab64-4c9b-b144-238e35ab3219/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.867305 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" path="/var/lib/kubelet/pods/310ed7ff-0788-436d-b811-a4aafc7f9f5f/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.868725 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4559a88d-6d07-44d0-a829-c35fd773fcfb" path="/var/lib/kubelet/pods/4559a88d-6d07-44d0-a829-c35fd773fcfb/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.878955 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" path="/var/lib/kubelet/pods/4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.880028 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73c57ace-65f0-40ab-996a-e5595851177a" path="/var/lib/kubelet/pods/73c57ace-65f0-40ab-996a-e5595851177a/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.880692 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75f196ee-9a0c-414f-a629-5b270c412c24" path="/var/lib/kubelet/pods/75f196ee-9a0c-414f-a629-5b270c412c24/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.886848 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77de7245-a8de-4912-b90f-669c57d8d563" path="/var/lib/kubelet/pods/77de7245-a8de-4912-b90f-669c57d8d563/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.887746 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" path="/var/lib/kubelet/pods/91afecb8-a07b-43e5-92e4-4e86cce7bcca/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.888415 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="989edff9-2d95-4839-9926-f5ee7dc65d19" path="/var/lib/kubelet/pods/989edff9-2d95-4839-9926-f5ee7dc65d19/volumes" Nov 28 11:47:26 crc kubenswrapper[4862]: I1128 11:47:26.891789 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca62671f-2a31-4980-93d9-2239defbfa57" path="/var/lib/kubelet/pods/ca62671f-2a31-4980-93d9-2239defbfa57/volumes" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.922240 4862 scope.go:117] "RemoveContainer" containerID="7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.923958 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d86c33d6-e42d-4544-af10-64f39f8201f7" path="/var/lib/kubelet/pods/d86c33d6-e42d-4544-af10-64f39f8201f7/volumes" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.926615 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8ef59ee-556c-4735-8cdf-e140aa608f86" path="/var/lib/kubelet/pods/e8ef59ee-556c-4735-8cdf-e140aa608f86/volumes" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.927584 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e949f4c6-f086-47e6-9a9d-b78f8b7837e7" path="/var/lib/kubelet/pods/e949f4c6-f086-47e6-9a9d-b78f8b7837e7/volumes" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.928173 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" path="/var/lib/kubelet/pods/edf3f6fa-2e74-489d-a32e-9faa5c0991ac/volumes" Nov 28 11:47:27 crc kubenswrapper[4862]: E1128 11:47:26.928458 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:27 crc kubenswrapper[4862]: E1128 11:47:26.928515 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts podName:ed3d880d-6762-4d57-881d-bc585318d500 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:30.928495975 +0000 UTC m=+1528.461009896 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts") pod "barbican8d34-account-delete-bdd2b" (UID: "ed3d880d-6762-4d57-881d-bc585318d500") : configmap "openstack-scripts" not found Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.929529 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f01191bd-12ff-43ea-9788-381a997a3c08" path="/var/lib/kubelet/pods/f01191bd-12ff-43ea-9788-381a997a3c08/volumes" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.930273 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc2928a-7858-4117-8786-723965b126f4" path="/var/lib/kubelet/pods/ffc2928a-7858-4117-8786-723965b126f4/volumes" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.930921 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.930944 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.930960 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7655db979f-pdkkm"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.930971 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7655db979f-pdkkm"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.930990 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:26.935479 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.132990 4862 scope.go:117] "RemoveContainer" containerID="94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980" Nov 28 11:47:27 crc kubenswrapper[4862]: E1128 11:47:27.133438 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980\": container with ID starting with 94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980 not found: ID does not exist" containerID="94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.133475 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980"} err="failed to get container status \"94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980\": rpc error: code = NotFound desc = could not find container \"94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980\": container with ID starting with 94f43d2f7bc47b057d00ce5ffb30c9a4c477f6a1c85a3618abaf3ed8bd96a980 not found: ID does not exist" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.133504 4862 scope.go:117] "RemoveContainer" containerID="7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755" Nov 28 11:47:27 crc kubenswrapper[4862]: E1128 11:47:27.133877 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755\": container with ID starting with 7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755 not found: ID does not exist" containerID="7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.133902 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755"} err="failed to get container status \"7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755\": rpc error: code = NotFound desc = could not find container \"7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755\": container with ID starting with 7885d6d8b6ccc4defc675a1017228348e0cf2949c3c5455c974b2a5b3e485755 not found: ID does not exist" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.133929 4862 scope.go:117] "RemoveContainer" containerID="0682b22558ac57d011862a3e8a94d6b4334e16c0f890d30fb6ae86e9faaac01e" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.184051 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.192711 4862 scope.go:117] "RemoveContainer" containerID="e487aac69ef596761ce6caee885f81d9e334146dc7a0a9ba8b24df61e5dc1e8f" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.219749 4862 scope.go:117] "RemoveContainer" containerID="5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233011 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233073 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233187 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-erlang-cookie\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233213 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-confd\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233251 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-server-conf\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233268 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-plugins-conf\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233305 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-pod-info\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233329 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-tls\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233386 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-plugins\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233425 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgkwp\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-kube-api-access-rgkwp\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.233463 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-erlang-cookie-secret\") pod \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\" (UID: \"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356\") " Nov 28 11:47:27 crc kubenswrapper[4862]: E1128 11:47:27.233833 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:27 crc kubenswrapper[4862]: E1128 11:47:27.233885 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts podName:880d930f-ea3a-4a7b-a437-5b164fbea09e nodeName:}" failed. No retries permitted until 2025-11-28 11:47:29.233870723 +0000 UTC m=+1526.766384644 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts") pod "cinder68e8-account-delete-rv4w4" (UID: "880d930f-ea3a-4a7b-a437-5b164fbea09e") : configmap "openstack-scripts" not found Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.234987 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.235235 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.236434 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.237694 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.238716 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.238759 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-kube-api-access-rgkwp" (OuterVolumeSpecName: "kube-api-access-rgkwp") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "kube-api-access-rgkwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.238750 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-pod-info" (OuterVolumeSpecName: "pod-info") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.240153 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.246051 4862 scope.go:117] "RemoveContainer" containerID="5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b" Nov 28 11:47:27 crc kubenswrapper[4862]: E1128 11:47:27.246715 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b\": container with ID starting with 5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b not found: ID does not exist" containerID="5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.246761 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b"} err="failed to get container status \"5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b\": rpc error: code = NotFound desc = could not find container \"5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b\": container with ID starting with 5bc1d43bee363cd32a0848868f76bcb352f41278c6ce1c6823ce8e09f68bde8b not found: ID does not exist" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.258461 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data" (OuterVolumeSpecName: "config-data") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.282279 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-server-conf" (OuterVolumeSpecName: "server-conf") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335115 4862 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-server-conf\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335142 4862 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335152 4862 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-pod-info\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335160 4862 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335169 4862 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335178 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgkwp\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-kube-api-access-rgkwp\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335187 4862 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335217 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335227 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.335236 4862 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.349559 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" (UID: "88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.351047 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.438722 4862 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.438760 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.461263 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5e37bdc4-788e-4889-9de7-682e77c3c5eb/ovn-northd/0.log" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.461331 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.500383 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540327 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-scripts\") pod \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540385 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-confd\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540458 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540506 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-config\") pod \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540539 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-plugins\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540562 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-server-conf\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540621 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-plugins-conf\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540665 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-rundir\") pod \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540692 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-northd-tls-certs\") pod \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540738 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-erlang-cookie\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540764 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540789 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stjnh\" (UniqueName: \"kubernetes.io/projected/5e37bdc4-788e-4889-9de7-682e77c3c5eb-kube-api-access-stjnh\") pod \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540819 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfwrl\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-kube-api-access-wfwrl\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540839 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-tls\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540880 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/527b2bce-1186-4186-8992-a3a63d2d2b22-pod-info\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540931 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-combined-ca-bundle\") pod \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540961 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/527b2bce-1186-4186-8992-a3a63d2d2b22-erlang-cookie-secret\") pod \"527b2bce-1186-4186-8992-a3a63d2d2b22\" (UID: \"527b2bce-1186-4186-8992-a3a63d2d2b22\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.540984 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-metrics-certs-tls-certs\") pod \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\" (UID: \"5e37bdc4-788e-4889-9de7-682e77c3c5eb\") " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.541222 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.541769 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.542579 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-scripts" (OuterVolumeSpecName: "scripts") pod "5e37bdc4-788e-4889-9de7-682e77c3c5eb" (UID: "5e37bdc4-788e-4889-9de7-682e77c3c5eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.544665 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.545350 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.546352 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "5e37bdc4-788e-4889-9de7-682e77c3c5eb" (UID: "5e37bdc4-788e-4889-9de7-682e77c3c5eb"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.546486 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-config" (OuterVolumeSpecName: "config") pod "5e37bdc4-788e-4889-9de7-682e77c3c5eb" (UID: "5e37bdc4-788e-4889-9de7-682e77c3c5eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.546707 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/527b2bce-1186-4186-8992-a3a63d2d2b22-pod-info" (OuterVolumeSpecName: "pod-info") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.546816 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e37bdc4-788e-4889-9de7-682e77c3c5eb-kube-api-access-stjnh" (OuterVolumeSpecName: "kube-api-access-stjnh") pod "5e37bdc4-788e-4889-9de7-682e77c3c5eb" (UID: "5e37bdc4-788e-4889-9de7-682e77c3c5eb"). InnerVolumeSpecName "kube-api-access-stjnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.547762 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.547820 4862 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.547836 4862 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.547850 4862 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.547894 4862 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.547924 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.547937 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stjnh\" (UniqueName: \"kubernetes.io/projected/5e37bdc4-788e-4889-9de7-682e77c3c5eb-kube-api-access-stjnh\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.547979 4862 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/527b2bce-1186-4186-8992-a3a63d2d2b22-pod-info\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.547994 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e37bdc4-788e-4889-9de7-682e77c3c5eb-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.549882 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-kube-api-access-wfwrl" (OuterVolumeSpecName: "kube-api-access-wfwrl") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "kube-api-access-wfwrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.553046 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.575849 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.581235 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527b2bce-1186-4186-8992-a3a63d2d2b22-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.605063 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data" (OuterVolumeSpecName: "config-data") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.622197 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e37bdc4-788e-4889-9de7-682e77c3c5eb" (UID: "5e37bdc4-788e-4889-9de7-682e77c3c5eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.649528 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.649562 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfwrl\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-kube-api-access-wfwrl\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.649572 4862 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.649581 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.649589 4862 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/527b2bce-1186-4186-8992-a3a63d2d2b22-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.649598 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.669787 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "5e37bdc4-788e-4889-9de7-682e77c3c5eb" (UID: "5e37bdc4-788e-4889-9de7-682e77c3c5eb"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.673749 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-server-conf" (OuterVolumeSpecName: "server-conf") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.687427 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "5e37bdc4-788e-4889-9de7-682e77c3c5eb" (UID: "5e37bdc4-788e-4889-9de7-682e77c3c5eb"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.708352 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "527b2bce-1186-4186-8992-a3a63d2d2b22" (UID: "527b2bce-1186-4186-8992-a3a63d2d2b22"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.751415 4862 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.751465 4862 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/527b2bce-1186-4186-8992-a3a63d2d2b22-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.751478 4862 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/527b2bce-1186-4186-8992-a3a63d2d2b22-server-conf\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.751490 4862 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e37bdc4-788e-4889-9de7-682e77c3c5eb-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.762825 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5e37bdc4-788e-4889-9de7-682e77c3c5eb/ovn-northd/0.log" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.762870 4862 generic.go:334] "Generic (PLEG): container finished" podID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerID="3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6" exitCode=139 Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.762940 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5e37bdc4-788e-4889-9de7-682e77c3c5eb","Type":"ContainerDied","Data":"3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6"} Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.762969 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5e37bdc4-788e-4889-9de7-682e77c3c5eb","Type":"ContainerDied","Data":"8d2a830fa44e976a4feb94ac2049af8d2b090df8bf4e0247a72e920f00732e06"} Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.762988 4862 scope.go:117] "RemoveContainer" containerID="23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.763133 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.774938 4862 generic.go:334] "Generic (PLEG): container finished" podID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" containerID="dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f" exitCode=0 Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.774998 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356","Type":"ContainerDied","Data":"dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f"} Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.775021 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356","Type":"ContainerDied","Data":"c4724491b275a3f9f3046ef2356616d2d16720ef4c64f64f138d4a6d71865a54"} Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.775077 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.783956 4862 generic.go:334] "Generic (PLEG): container finished" podID="527b2bce-1186-4186-8992-a3a63d2d2b22" containerID="d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97" exitCode=0 Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.784017 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"527b2bce-1186-4186-8992-a3a63d2d2b22","Type":"ContainerDied","Data":"d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97"} Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.784043 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"527b2bce-1186-4186-8992-a3a63d2d2b22","Type":"ContainerDied","Data":"bbf72032ab897813bfd4f4fed577a12e1dbbf694c3bfa01700bfc71a8e549018"} Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.784042 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.797805 4862 scope.go:117] "RemoveContainer" containerID="3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.819438 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.854145 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.871788 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.875059 4862 scope.go:117] "RemoveContainer" containerID="23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.877146 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.881900 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 11:47:27 crc kubenswrapper[4862]: E1128 11:47:27.882536 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3\": container with ID starting with 23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3 not found: ID does not exist" containerID="23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.882572 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3"} err="failed to get container status \"23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3\": rpc error: code = NotFound desc = could not find container \"23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3\": container with ID starting with 23d45a75a7d4c79e57f75bc3ff6c9f8b41e2af657e2aed3c9df417a492d41bd3 not found: ID does not exist" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.882594 4862 scope.go:117] "RemoveContainer" containerID="3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6" Nov 28 11:47:27 crc kubenswrapper[4862]: E1128 11:47:27.883027 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6\": container with ID starting with 3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6 not found: ID does not exist" containerID="3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.883077 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6"} err="failed to get container status \"3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6\": rpc error: code = NotFound desc = could not find container \"3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6\": container with ID starting with 3b3cd001d45ac12584cf20a7be9b7b53e9c3a7060d6ebd9a0b0527cc886a63f6 not found: ID does not exist" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.883127 4862 scope.go:117] "RemoveContainer" containerID="dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f" Nov 28 11:47:27 crc kubenswrapper[4862]: I1128 11:47:27.886608 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.088501 4862 scope.go:117] "RemoveContainer" containerID="28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.117828 4862 scope.go:117] "RemoveContainer" containerID="dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f" Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.118299 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f\": container with ID starting with dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f not found: ID does not exist" containerID="dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.118344 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f"} err="failed to get container status \"dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f\": rpc error: code = NotFound desc = could not find container \"dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f\": container with ID starting with dc77c997a2a596646bfe80ad4473413bb2e7954ecfcae5ef849e71f93b265d8f not found: ID does not exist" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.118364 4862 scope.go:117] "RemoveContainer" containerID="28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245" Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.119257 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245\": container with ID starting with 28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245 not found: ID does not exist" containerID="28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.119299 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245"} err="failed to get container status \"28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245\": rpc error: code = NotFound desc = could not find container \"28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245\": container with ID starting with 28c752a7540475638f4496afa6aa998b6899fb959db0d3f8544e5fc3884f5245 not found: ID does not exist" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.119328 4862 scope.go:117] "RemoveContainer" containerID="d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.141408 4862 scope.go:117] "RemoveContainer" containerID="b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.194483 4862 scope.go:117] "RemoveContainer" containerID="d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97" Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.194812 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97\": container with ID starting with d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97 not found: ID does not exist" containerID="d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.194852 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97"} err="failed to get container status \"d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97\": rpc error: code = NotFound desc = could not find container \"d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97\": container with ID starting with d9d313fba57b829c6e94c8920166ae70be4156c783e9ea2a7cdbbfd6467ccb97 not found: ID does not exist" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.194882 4862 scope.go:117] "RemoveContainer" containerID="b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2" Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.195149 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2\": container with ID starting with b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2 not found: ID does not exist" containerID="b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.195166 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2"} err="failed to get container status \"b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2\": rpc error: code = NotFound desc = could not find container \"b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2\": container with ID starting with b9e3bd454f5b049f79f43e6f578fd6d95d6eb672980cf0e38d6655f5348e9dd2 not found: ID does not exist" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.238450 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.257401 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-logs\") pod \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.257599 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-combined-ca-bundle\") pod \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.257662 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data\") pod \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.257692 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb5sl\" (UniqueName: \"kubernetes.io/projected/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-kube-api-access-hb5sl\") pod \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.257753 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data-custom\") pod \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\" (UID: \"e4f23c53-e6f8-4aca-a1e4-969a58b70a38\") " Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.258323 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.258397 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts podName:dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:47:32.258362721 +0000 UTC m=+1529.790876652 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts") pod "neutronc4cb-account-delete-9jrbn" (UID: "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f") : configmap "openstack-scripts" not found Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.261745 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-logs" (OuterVolumeSpecName: "logs") pod "e4f23c53-e6f8-4aca-a1e4-969a58b70a38" (UID: "e4f23c53-e6f8-4aca-a1e4-969a58b70a38"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.265692 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e4f23c53-e6f8-4aca-a1e4-969a58b70a38" (UID: "e4f23c53-e6f8-4aca-a1e4-969a58b70a38"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.266270 4862 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 28 11:47:28 crc kubenswrapper[4862]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-11-28T11:47:21Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 28 11:47:28 crc kubenswrapper[4862]: /etc/init.d/functions: line 589: 410 Alarm clock "$@" Nov 28 11:47:28 crc kubenswrapper[4862]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-hkgwr" message=< Nov 28 11:47:28 crc kubenswrapper[4862]: Exiting ovn-controller (1) [FAILED] Nov 28 11:47:28 crc kubenswrapper[4862]: Killing ovn-controller (1) [ OK ] Nov 28 11:47:28 crc kubenswrapper[4862]: Killing ovn-controller (1) with SIGKILL [ OK ] Nov 28 11:47:28 crc kubenswrapper[4862]: 2025-11-28T11:47:21Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 28 11:47:28 crc kubenswrapper[4862]: /etc/init.d/functions: line 589: 410 Alarm clock "$@" Nov 28 11:47:28 crc kubenswrapper[4862]: > Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.266309 4862 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 28 11:47:28 crc kubenswrapper[4862]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-11-28T11:47:21Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 28 11:47:28 crc kubenswrapper[4862]: /etc/init.d/functions: line 589: 410 Alarm clock "$@" Nov 28 11:47:28 crc kubenswrapper[4862]: > pod="openstack/ovn-controller-hkgwr" podUID="b1bdf176-776a-4211-8b43-ff90225de25a" containerName="ovn-controller" containerID="cri-o://4f561afbdd4fb1259076c496dea33b5413f902ebee541fb48eb7e1aa8fa45027" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.266365 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-hkgwr" podUID="b1bdf176-776a-4211-8b43-ff90225de25a" containerName="ovn-controller" containerID="cri-o://4f561afbdd4fb1259076c496dea33b5413f902ebee541fb48eb7e1aa8fa45027" gracePeriod=22 Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.281104 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-kube-api-access-hb5sl" (OuterVolumeSpecName: "kube-api-access-hb5sl") pod "e4f23c53-e6f8-4aca-a1e4-969a58b70a38" (UID: "e4f23c53-e6f8-4aca-a1e4-969a58b70a38"). InnerVolumeSpecName "kube-api-access-hb5sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.310402 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4f23c53-e6f8-4aca-a1e4-969a58b70a38" (UID: "e4f23c53-e6f8-4aca-a1e4-969a58b70a38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.336878 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data" (OuterVolumeSpecName: "config-data") pod "e4f23c53-e6f8-4aca-a1e4-969a58b70a38" (UID: "e4f23c53-e6f8-4aca-a1e4-969a58b70a38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.360300 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.360344 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.360363 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb5sl\" (UniqueName: \"kubernetes.io/projected/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-kube-api-access-hb5sl\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.360380 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.360414 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4f23c53-e6f8-4aca-a1e4-969a58b70a38-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.484047 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.494292 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664009 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-ceilometer-tls-certs\") pod \"6b1b592a-d848-4857-894c-8f05a736dc6d\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664116 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-combined-ca-bundle\") pod \"6b1b592a-d848-4857-894c-8f05a736dc6d\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664173 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5203b21-897c-4331-a878-1d5c949a3e75-logs\") pod \"d5203b21-897c-4331-a878-1d5c949a3e75\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664207 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-scripts\") pod \"6b1b592a-d848-4857-894c-8f05a736dc6d\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664263 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkdk8\" (UniqueName: \"kubernetes.io/projected/d5203b21-897c-4331-a878-1d5c949a3e75-kube-api-access-dkdk8\") pod \"d5203b21-897c-4331-a878-1d5c949a3e75\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664290 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-sg-core-conf-yaml\") pod \"6b1b592a-d848-4857-894c-8f05a736dc6d\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664353 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-config-data\") pod \"6b1b592a-d848-4857-894c-8f05a736dc6d\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664411 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-run-httpd\") pod \"6b1b592a-d848-4857-894c-8f05a736dc6d\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664483 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-combined-ca-bundle\") pod \"d5203b21-897c-4331-a878-1d5c949a3e75\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664529 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data\") pod \"d5203b21-897c-4331-a878-1d5c949a3e75\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664599 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data-custom\") pod \"d5203b21-897c-4331-a878-1d5c949a3e75\" (UID: \"d5203b21-897c-4331-a878-1d5c949a3e75\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664666 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-log-httpd\") pod \"6b1b592a-d848-4857-894c-8f05a736dc6d\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.664723 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6gsx\" (UniqueName: \"kubernetes.io/projected/6b1b592a-d848-4857-894c-8f05a736dc6d-kube-api-access-x6gsx\") pod \"6b1b592a-d848-4857-894c-8f05a736dc6d\" (UID: \"6b1b592a-d848-4857-894c-8f05a736dc6d\") " Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.665758 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6b1b592a-d848-4857-894c-8f05a736dc6d" (UID: "6b1b592a-d848-4857-894c-8f05a736dc6d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.666689 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5203b21-897c-4331-a878-1d5c949a3e75-logs" (OuterVolumeSpecName: "logs") pod "d5203b21-897c-4331-a878-1d5c949a3e75" (UID: "d5203b21-897c-4331-a878-1d5c949a3e75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.670262 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5203b21-897c-4331-a878-1d5c949a3e75-kube-api-access-dkdk8" (OuterVolumeSpecName: "kube-api-access-dkdk8") pod "d5203b21-897c-4331-a878-1d5c949a3e75" (UID: "d5203b21-897c-4331-a878-1d5c949a3e75"). InnerVolumeSpecName "kube-api-access-dkdk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.671389 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6b1b592a-d848-4857-894c-8f05a736dc6d" (UID: "6b1b592a-d848-4857-894c-8f05a736dc6d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.672344 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b1b592a-d848-4857-894c-8f05a736dc6d-kube-api-access-x6gsx" (OuterVolumeSpecName: "kube-api-access-x6gsx") pod "6b1b592a-d848-4857-894c-8f05a736dc6d" (UID: "6b1b592a-d848-4857-894c-8f05a736dc6d"). InnerVolumeSpecName "kube-api-access-x6gsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.684187 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d5203b21-897c-4331-a878-1d5c949a3e75" (UID: "d5203b21-897c-4331-a878-1d5c949a3e75"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.693137 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-scripts" (OuterVolumeSpecName: "scripts") pod "6b1b592a-d848-4857-894c-8f05a736dc6d" (UID: "6b1b592a-d848-4857-894c-8f05a736dc6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.696559 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5203b21-897c-4331-a878-1d5c949a3e75" (UID: "d5203b21-897c-4331-a878-1d5c949a3e75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.719400 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "6b1b592a-d848-4857-894c-8f05a736dc6d" (UID: "6b1b592a-d848-4857-894c-8f05a736dc6d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.720115 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6b1b592a-d848-4857-894c-8f05a736dc6d" (UID: "6b1b592a-d848-4857-894c-8f05a736dc6d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.728441 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data" (OuterVolumeSpecName: "config-data") pod "d5203b21-897c-4331-a878-1d5c949a3e75" (UID: "d5203b21-897c-4331-a878-1d5c949a3e75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.749550 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b1b592a-d848-4857-894c-8f05a736dc6d" (UID: "6b1b592a-d848-4857-894c-8f05a736dc6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771114 4862 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771148 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771158 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771167 4862 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d5203b21-897c-4331-a878-1d5c949a3e75-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771178 4862 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b1b592a-d848-4857-894c-8f05a736dc6d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771188 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6gsx\" (UniqueName: \"kubernetes.io/projected/6b1b592a-d848-4857-894c-8f05a736dc6d-kube-api-access-x6gsx\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771197 4862 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771205 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771213 4862 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5203b21-897c-4331-a878-1d5c949a3e75-logs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771221 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771229 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkdk8\" (UniqueName: \"kubernetes.io/projected/d5203b21-897c-4331-a878-1d5c949a3e75-kube-api-access-dkdk8\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.771239 4862 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.807298 4862 generic.go:334] "Generic (PLEG): container finished" podID="d5203b21-897c-4331-a878-1d5c949a3e75" containerID="31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7" exitCode=0 Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.807518 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6df5c754df-6lvmr" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.807839 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6df5c754df-6lvmr" event={"ID":"d5203b21-897c-4331-a878-1d5c949a3e75","Type":"ContainerDied","Data":"31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7"} Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.807886 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6df5c754df-6lvmr" event={"ID":"d5203b21-897c-4331-a878-1d5c949a3e75","Type":"ContainerDied","Data":"749f84f1d79e4dc9a76a2699b9668ed8f2dee3349b07cfde631e2b369b80b82f"} Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.807903 4862 scope.go:117] "RemoveContainer" containerID="31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.813695 4862 generic.go:334] "Generic (PLEG): container finished" podID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerID="f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe" exitCode=0 Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.813789 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerDied","Data":"f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe"} Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.813826 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b1b592a-d848-4857-894c-8f05a736dc6d","Type":"ContainerDied","Data":"edcb718a300c14df2b671048e5fbc7738c2ac9e67def205db08d60b1a0934fed"} Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.814051 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.816560 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-hkgwr_b1bdf176-776a-4211-8b43-ff90225de25a/ovn-controller/0.log" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.816781 4862 generic.go:334] "Generic (PLEG): container finished" podID="b1bdf176-776a-4211-8b43-ff90225de25a" containerID="4f561afbdd4fb1259076c496dea33b5413f902ebee541fb48eb7e1aa8fa45027" exitCode=137 Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.816916 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hkgwr" event={"ID":"b1bdf176-776a-4211-8b43-ff90225de25a","Type":"ContainerDied","Data":"4f561afbdd4fb1259076c496dea33b5413f902ebee541fb48eb7e1aa8fa45027"} Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.817016 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hkgwr" event={"ID":"b1bdf176-776a-4211-8b43-ff90225de25a","Type":"ContainerDied","Data":"ff2430e6a4891c2e18501cf0209ccc7e522f0465b6639bd7b47c8dde544ce99a"} Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.817103 4862 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff2430e6a4891c2e18501cf0209ccc7e522f0465b6639bd7b47c8dde544ce99a" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.822941 4862 generic.go:334] "Generic (PLEG): container finished" podID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerID="b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6" exitCode=0 Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.823177 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" event={"ID":"e4f23c53-e6f8-4aca-a1e4-969a58b70a38","Type":"ContainerDied","Data":"b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6"} Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.823279 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" event={"ID":"e4f23c53-e6f8-4aca-a1e4-969a58b70a38","Type":"ContainerDied","Data":"048866d2590e7fca7d149171a3413cf0ef36cd649289e039311ed45d486fc1fa"} Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.823429 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7756445dd6-vwrsv" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.841639 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-config-data" (OuterVolumeSpecName: "config-data") pod "6b1b592a-d848-4857-894c-8f05a736dc6d" (UID: "6b1b592a-d848-4857-894c-8f05a736dc6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.864245 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.162:8776/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.866103 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c30075d-2cd8-4f5a-9168-33c55459f636" path="/var/lib/kubelet/pods/0c30075d-2cd8-4f5a-9168-33c55459f636/volumes" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.866978 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="527b2bce-1186-4186-8992-a3a63d2d2b22" path="/var/lib/kubelet/pods/527b2bce-1186-4186-8992-a3a63d2d2b22/volumes" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.867787 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" path="/var/lib/kubelet/pods/5e37bdc4-788e-4889-9de7-682e77c3c5eb/volumes" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.869288 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" path="/var/lib/kubelet/pods/88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356/volumes" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.869942 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a824cc2c-2812-4d91-9e9f-9736a27fe794" path="/var/lib/kubelet/pods/a824cc2c-2812-4d91-9e9f-9736a27fe794/volumes" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.871087 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae02e889-f98f-4678-b15f-d91f711083f9" path="/var/lib/kubelet/pods/ae02e889-f98f-4678-b15f-d91f711083f9/volumes" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.872263 4862 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b1b592a-d848-4857-894c-8f05a736dc6d-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.900286 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-hkgwr_b1bdf176-776a-4211-8b43-ff90225de25a/ovn-controller/0.log" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.900482 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hkgwr" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.912051 4862 scope.go:117] "RemoveContainer" containerID="cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.943575 4862 scope.go:117] "RemoveContainer" containerID="31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7" Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.944229 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7\": container with ID starting with 31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7 not found: ID does not exist" containerID="31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.944279 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7"} err="failed to get container status \"31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7\": rpc error: code = NotFound desc = could not find container \"31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7\": container with ID starting with 31cbdc1add7b5872dc7b18a25c5e65db8892135d83fa8e408be2a120804ccef7 not found: ID does not exist" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.944310 4862 scope.go:117] "RemoveContainer" containerID="cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d" Nov 28 11:47:28 crc kubenswrapper[4862]: E1128 11:47:28.944866 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d\": container with ID starting with cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d not found: ID does not exist" containerID="cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.944928 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d"} err="failed to get container status \"cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d\": rpc error: code = NotFound desc = could not find container \"cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d\": container with ID starting with cab0d065bd1fe54f64ae027132b2aa2c7d8ff5b5555aedf09b8a66e58b56a24d not found: ID does not exist" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.944964 4862 scope.go:117] "RemoveContainer" containerID="35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde" Nov 28 11:47:28 crc kubenswrapper[4862]: I1128 11:47:28.945984 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6df5c754df-6lvmr"] Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:28.994411 4862 scope.go:117] "RemoveContainer" containerID="eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.026926 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-6df5c754df-6lvmr"] Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.036481 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5ps8s"] Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037466 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="openstack-network-exporter" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037493 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="openstack-network-exporter" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037527 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c30075d-2cd8-4f5a-9168-33c55459f636" containerName="keystone-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037536 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c30075d-2cd8-4f5a-9168-33c55459f636" containerName="keystone-api" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037557 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="164085a5-ab64-4c9b-b144-238e35ab3219" containerName="kube-state-metrics" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037571 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="164085a5-ab64-4c9b-b144-238e35ab3219" containerName="kube-state-metrics" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037590 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037603 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037622 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae02e889-f98f-4678-b15f-d91f711083f9" containerName="nova-cell1-conductor-conductor" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037631 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae02e889-f98f-4678-b15f-d91f711083f9" containerName="nova-cell1-conductor-conductor" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037652 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037660 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-api" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037679 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f196ee-9a0c-414f-a629-5b270c412c24" containerName="mariadb-account-delete" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037687 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f196ee-9a0c-414f-a629-5b270c412c24" containerName="mariadb-account-delete" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037701 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f01191bd-12ff-43ea-9788-381a997a3c08" containerName="galera" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037710 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f01191bd-12ff-43ea-9788-381a997a3c08" containerName="galera" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037734 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerName="glance-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037744 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerName="glance-log" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037769 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerName="proxy-server" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037776 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerName="proxy-server" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037792 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037800 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api-log" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037817 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527b2bce-1186-4186-8992-a3a63d2d2b22" containerName="rabbitmq" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037825 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="527b2bce-1186-4186-8992-a3a63d2d2b22" containerName="rabbitmq" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037848 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerName="proxy-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037855 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerName="proxy-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037865 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerName="glance-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037872 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerName="glance-log" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037894 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1bdf176-776a-4211-8b43-ff90225de25a" containerName="ovn-controller" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037903 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1bdf176-776a-4211-8b43-ff90225de25a" containerName="ovn-controller" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037922 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" containerName="setup-container" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037930 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" containerName="setup-container" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037947 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerName="cinder-scheduler" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037960 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerName="cinder-scheduler" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.037985 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerName="glance-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.037993 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerName="glance-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.038016 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca62671f-2a31-4980-93d9-2239defbfa57" containerName="placement-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.038024 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca62671f-2a31-4980-93d9-2239defbfa57" containerName="placement-api" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.038032 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01f7b6ff-ee76-4961-9baf-930743da6143" containerName="nova-cell0-conductor-conductor" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.038039 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="01f7b6ff-ee76-4961-9baf-930743da6143" containerName="nova-cell0-conductor-conductor" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.038056 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.038063 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.038083 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="proxy-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.038111 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="proxy-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.038126 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527b2bce-1186-4186-8992-a3a63d2d2b22" containerName="setup-container" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.038134 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="527b2bce-1186-4186-8992-a3a63d2d2b22" containerName="setup-container" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.038151 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerName="glance-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.038158 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerName="glance-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.038170 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="ceilometer-central-agent" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.038178 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="ceilometer-central-agent" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.038634 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="sg-core" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.040299 4862 scope.go:117] "RemoveContainer" containerID="f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.041896 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="sg-core" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.041949 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-metadata" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.041961 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-metadata" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042025 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4559a88d-6d07-44d0-a829-c35fd773fcfb" containerName="memcached" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042040 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="4559a88d-6d07-44d0-a829-c35fd773fcfb" containerName="memcached" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042060 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca62671f-2a31-4980-93d9-2239defbfa57" containerName="placement-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042068 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca62671f-2a31-4980-93d9-2239defbfa57" containerName="placement-log" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042104 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerName="barbican-keystone-listener" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042117 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerName="barbican-keystone-listener" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042144 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="ovn-northd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042152 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="ovn-northd" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042166 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a824cc2c-2812-4d91-9e9f-9736a27fe794" containerName="nova-scheduler-scheduler" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042174 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="a824cc2c-2812-4d91-9e9f-9736a27fe794" containerName="nova-scheduler-scheduler" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042203 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f01191bd-12ff-43ea-9788-381a997a3c08" containerName="mysql-bootstrap" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042211 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f01191bd-12ff-43ea-9788-381a997a3c08" containerName="mysql-bootstrap" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042230 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042237 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-log" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042250 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5203b21-897c-4331-a878-1d5c949a3e75" containerName="barbican-worker" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042259 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5203b21-897c-4331-a878-1d5c949a3e75" containerName="barbican-worker" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042272 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerName="probe" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042281 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerName="probe" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042296 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerName="barbican-keystone-listener-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042304 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerName="barbican-keystone-listener-log" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042331 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5203b21-897c-4331-a878-1d5c949a3e75" containerName="barbican-worker-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042340 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5203b21-897c-4331-a878-1d5c949a3e75" containerName="barbican-worker-log" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042355 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" containerName="rabbitmq" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042363 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" containerName="rabbitmq" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042382 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042394 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api-log" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042418 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e949f4c6-f086-47e6-9a9d-b78f8b7837e7" containerName="mariadb-account-delete" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042426 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e949f4c6-f086-47e6-9a9d-b78f8b7837e7" containerName="mariadb-account-delete" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042452 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="ceilometer-notification-agent" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042460 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="ceilometer-notification-agent" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.042478 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042486 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042914 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="ceilometer-notification-agent" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042936 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="proxy-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042958 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042967 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="4559a88d-6d07-44d0-a829-c35fd773fcfb" containerName="memcached" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.042990 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="sg-core" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043012 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="164085a5-ab64-4c9b-b144-238e35ab3219" containerName="kube-state-metrics" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043037 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f196ee-9a0c-414f-a629-5b270c412c24" containerName="mariadb-account-delete" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043064 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="91afecb8-a07b-43e5-92e4-4e86cce7bcca" containerName="barbican-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043562 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerName="glance-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043581 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e949f4c6-f086-47e6-9a9d-b78f8b7837e7" containerName="mariadb-account-delete" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043598 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerName="barbican-keystone-listener-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043609 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1bdf176-776a-4211-8b43-ff90225de25a" containerName="ovn-controller" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043635 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerName="glance-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043656 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerName="probe" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043665 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043680 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043706 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="310ed7ff-0788-436d-b811-a4aafc7f9f5f" containerName="nova-metadata-metadata" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043722 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="openstack-network-exporter" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043743 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc2928a-7858-4117-8786-723965b126f4" containerName="cinder-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043813 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f01191bd-12ff-43ea-9788-381a997a3c08" containerName="galera" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043837 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e37bdc4-788e-4889-9de7-682e77c3c5eb" containerName="ovn-northd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043848 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae02e889-f98f-4678-b15f-d91f711083f9" containerName="nova-cell1-conductor-conductor" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043870 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="527b2bce-1186-4186-8992-a3a63d2d2b22" containerName="rabbitmq" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043888 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="a824cc2c-2812-4d91-9e9f-9736a27fe794" containerName="nova-scheduler-scheduler" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043903 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="989edff9-2d95-4839-9926-f5ee7dc65d19" containerName="cinder-scheduler" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043918 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" containerName="ceilometer-central-agent" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043935 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5203b21-897c-4331-a878-1d5c949a3e75" containerName="barbican-worker-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043958 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerName="proxy-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.043969 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b7fd5f-d4cc-4bd9-b9ec-039cf4a25356" containerName="rabbitmq" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044000 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca62671f-2a31-4980-93d9-2239defbfa57" containerName="placement-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044016 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca62671f-2a31-4980-93d9-2239defbfa57" containerName="placement-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044039 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c30075d-2cd8-4f5a-9168-33c55459f636" containerName="keystone-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044064 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ef59ee-556c-4735-8cdf-e140aa608f86" containerName="glance-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044075 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-api" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044115 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="77de7245-a8de-4912-b90f-669c57d8d563" containerName="nova-api-log" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044130 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="01f7b6ff-ee76-4961-9baf-930743da6143" containerName="nova-cell0-conductor-conductor" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044147 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5203b21-897c-4331-a878-1d5c949a3e75" containerName="barbican-worker" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044164 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf3f6fa-2e74-489d-a32e-9faa5c0991ac" containerName="proxy-server" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044179 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" containerName="barbican-keystone-listener" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.044198 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fcf5ac0-acbd-422a-8c6a-9cc34ca43d8a" containerName="glance-httpd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.047827 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.054364 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7756445dd6-vwrsv"] Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.062860 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7756445dd6-vwrsv"] Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.073337 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5ps8s"] Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.076970 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run-ovn\") pod \"b1bdf176-776a-4211-8b43-ff90225de25a\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.077124 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1bdf176-776a-4211-8b43-ff90225de25a-scripts\") pod \"b1bdf176-776a-4211-8b43-ff90225de25a\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.077205 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-combined-ca-bundle\") pod \"b1bdf176-776a-4211-8b43-ff90225de25a\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.077312 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-ovn-controller-tls-certs\") pod \"b1bdf176-776a-4211-8b43-ff90225de25a\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.077405 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8skk\" (UniqueName: \"kubernetes.io/projected/b1bdf176-776a-4211-8b43-ff90225de25a-kube-api-access-t8skk\") pod \"b1bdf176-776a-4211-8b43-ff90225de25a\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.077503 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run\") pod \"b1bdf176-776a-4211-8b43-ff90225de25a\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.077603 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-log-ovn\") pod \"b1bdf176-776a-4211-8b43-ff90225de25a\" (UID: \"b1bdf176-776a-4211-8b43-ff90225de25a\") " Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.077989 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b1bdf176-776a-4211-8b43-ff90225de25a" (UID: "b1bdf176-776a-4211-8b43-ff90225de25a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.078075 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b1bdf176-776a-4211-8b43-ff90225de25a" (UID: "b1bdf176-776a-4211-8b43-ff90225de25a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.085933 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1bdf176-776a-4211-8b43-ff90225de25a-scripts" (OuterVolumeSpecName: "scripts") pod "b1bdf176-776a-4211-8b43-ff90225de25a" (UID: "b1bdf176-776a-4211-8b43-ff90225de25a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.087878 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run" (OuterVolumeSpecName: "var-run") pod "b1bdf176-776a-4211-8b43-ff90225de25a" (UID: "b1bdf176-776a-4211-8b43-ff90225de25a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.099465 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1bdf176-776a-4211-8b43-ff90225de25a-kube-api-access-t8skk" (OuterVolumeSpecName: "kube-api-access-t8skk") pod "b1bdf176-776a-4211-8b43-ff90225de25a" (UID: "b1bdf176-776a-4211-8b43-ff90225de25a"). InnerVolumeSpecName "kube-api-access-t8skk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.108474 4862 scope.go:117] "RemoveContainer" containerID="11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.111396 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1bdf176-776a-4211-8b43-ff90225de25a" (UID: "b1bdf176-776a-4211-8b43-ff90225de25a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.157022 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "b1bdf176-776a-4211-8b43-ff90225de25a" (UID: "b1bdf176-776a-4211-8b43-ff90225de25a"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.160338 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.161612 4862 scope.go:117] "RemoveContainer" containerID="35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.163481 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde\": container with ID starting with 35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde not found: ID does not exist" containerID="35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.163559 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde"} err="failed to get container status \"35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde\": rpc error: code = NotFound desc = could not find container \"35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde\": container with ID starting with 35cadef861398bcede9e428565c73dc8c031909fc02bd37d60416119c7101bde not found: ID does not exist" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.163585 4862 scope.go:117] "RemoveContainer" containerID="eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.163817 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee\": container with ID starting with eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee not found: ID does not exist" containerID="eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.163837 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee"} err="failed to get container status \"eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee\": rpc error: code = NotFound desc = could not find container \"eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee\": container with ID starting with eddc94df3078bd29deb2a2d41c60daf140dbd900ac0cd72e8d788fa1dae46bee not found: ID does not exist" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.163851 4862 scope.go:117] "RemoveContainer" containerID="f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.164014 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe\": container with ID starting with f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe not found: ID does not exist" containerID="f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.164031 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe"} err="failed to get container status \"f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe\": rpc error: code = NotFound desc = could not find container \"f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe\": container with ID starting with f3e634c1c41ded0f5ff55b432f72271e3319571f68cb3709514e84130f1d25fe not found: ID does not exist" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.164044 4862 scope.go:117] "RemoveContainer" containerID="11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.164216 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945\": container with ID starting with 11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945 not found: ID does not exist" containerID="11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.164231 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945"} err="failed to get container status \"11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945\": rpc error: code = NotFound desc = could not find container \"11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945\": container with ID starting with 11c8bdc43810cea6c16901d8ddd4e0b8bbf46011f475dd2c4b47c519275bc945 not found: ID does not exist" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.164242 4862 scope.go:117] "RemoveContainer" containerID="b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.178791 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-catalog-content\") pod \"community-operators-5ps8s\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.178950 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-utilities\") pod \"community-operators-5ps8s\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.179044 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrwb6\" (UniqueName: \"kubernetes.io/projected/8631bc3d-1854-4bfa-b14a-e68dca7567e5-kube-api-access-rrwb6\") pod \"community-operators-5ps8s\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.179124 4862 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.179136 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1bdf176-776a-4211-8b43-ff90225de25a-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.179145 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.179155 4862 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1bdf176-776a-4211-8b43-ff90225de25a-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.179166 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8skk\" (UniqueName: \"kubernetes.io/projected/b1bdf176-776a-4211-8b43-ff90225de25a-kube-api-access-t8skk\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.179175 4862 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.179183 4862 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b1bdf176-776a-4211-8b43-ff90225de25a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.184053 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.190418 4862 scope.go:117] "RemoveContainer" containerID="d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.210355 4862 scope.go:117] "RemoveContainer" containerID="b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.210946 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6\": container with ID starting with b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6 not found: ID does not exist" containerID="b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.210984 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6"} err="failed to get container status \"b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6\": rpc error: code = NotFound desc = could not find container \"b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6\": container with ID starting with b4b5b1ed4e7cf61c55566758b15e494d8e38f1f3abb9619c7170f0442ecfd1c6 not found: ID does not exist" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.211011 4862 scope.go:117] "RemoveContainer" containerID="d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.211538 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5\": container with ID starting with d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5 not found: ID does not exist" containerID="d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.211558 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5"} err="failed to get container status \"d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5\": rpc error: code = NotFound desc = could not find container \"d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5\": container with ID starting with d54a14f973b991fc4e50cef38b2bc0703ed4271aa4bcba560a20ca0d8a52d9d5 not found: ID does not exist" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.234310 4862 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="4559a88d-6d07-44d0-a829-c35fd773fcfb" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.105:11211: i/o timeout" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.280827 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrwb6\" (UniqueName: \"kubernetes.io/projected/8631bc3d-1854-4bfa-b14a-e68dca7567e5-kube-api-access-rrwb6\") pod \"community-operators-5ps8s\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.280934 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-catalog-content\") pod \"community-operators-5ps8s\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.280990 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-utilities\") pod \"community-operators-5ps8s\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.281182 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.281257 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts podName:880d930f-ea3a-4a7b-a437-5b164fbea09e nodeName:}" failed. No retries permitted until 2025-11-28 11:47:33.281233383 +0000 UTC m=+1530.813747304 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts") pod "cinder68e8-account-delete-rv4w4" (UID: "880d930f-ea3a-4a7b-a437-5b164fbea09e") : configmap "openstack-scripts" not found Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.281646 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-utilities\") pod \"community-operators-5ps8s\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.281978 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-catalog-content\") pod \"community-operators-5ps8s\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.298662 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrwb6\" (UniqueName: \"kubernetes.io/projected/8631bc3d-1854-4bfa-b14a-e68dca7567e5-kube-api-access-rrwb6\") pod \"community-operators-5ps8s\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.374024 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.836046 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hkgwr" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.871445 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hkgwr"] Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.877269 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-hkgwr"] Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.914923 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.915423 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.915927 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.915988 4862 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.916257 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.920422 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.922441 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:29 crc kubenswrapper[4862]: E1128 11:47:29.922494 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" Nov 28 11:47:29 crc kubenswrapper[4862]: I1128 11:47:29.940610 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5ps8s"] Nov 28 11:47:30 crc kubenswrapper[4862]: I1128 11:47:30.846640 4862 generic.go:334] "Generic (PLEG): container finished" podID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerID="34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3" exitCode=0 Nov 28 11:47:30 crc kubenswrapper[4862]: I1128 11:47:30.847021 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b1b592a-d848-4857-894c-8f05a736dc6d" path="/var/lib/kubelet/pods/6b1b592a-d848-4857-894c-8f05a736dc6d/volumes" Nov 28 11:47:30 crc kubenswrapper[4862]: I1128 11:47:30.848074 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1bdf176-776a-4211-8b43-ff90225de25a" path="/var/lib/kubelet/pods/b1bdf176-776a-4211-8b43-ff90225de25a/volumes" Nov 28 11:47:30 crc kubenswrapper[4862]: I1128 11:47:30.848723 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5203b21-897c-4331-a878-1d5c949a3e75" path="/var/lib/kubelet/pods/d5203b21-897c-4331-a878-1d5c949a3e75/volumes" Nov 28 11:47:30 crc kubenswrapper[4862]: I1128 11:47:30.849926 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f23c53-e6f8-4aca-a1e4-969a58b70a38" path="/var/lib/kubelet/pods/e4f23c53-e6f8-4aca-a1e4-969a58b70a38/volumes" Nov 28 11:47:30 crc kubenswrapper[4862]: I1128 11:47:30.850590 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ps8s" event={"ID":"8631bc3d-1854-4bfa-b14a-e68dca7567e5","Type":"ContainerDied","Data":"34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3"} Nov 28 11:47:30 crc kubenswrapper[4862]: I1128 11:47:30.850651 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ps8s" event={"ID":"8631bc3d-1854-4bfa-b14a-e68dca7567e5","Type":"ContainerStarted","Data":"21bcebc1a07581ebaf8ecfea5d759ecb32c5abd1d006d78834c9d9148e1f57ea"} Nov 28 11:47:30 crc kubenswrapper[4862]: E1128 11:47:30.908851 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:30 crc kubenswrapper[4862]: E1128 11:47:30.908912 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts podName:409c7014-982c-4c60-b374-0ff46c398db4 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:38.908897711 +0000 UTC m=+1536.441411622 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts") pod "novaapi144c-account-delete-ls9c4" (UID: "409c7014-982c-4c60-b374-0ff46c398db4") : configmap "openstack-scripts" not found Nov 28 11:47:31 crc kubenswrapper[4862]: E1128 11:47:31.010577 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:31 crc kubenswrapper[4862]: E1128 11:47:31.010691 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts podName:ed3d880d-6762-4d57-881d-bc585318d500 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:39.010659127 +0000 UTC m=+1536.543173108 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts") pod "barbican8d34-account-delete-bdd2b" (UID: "ed3d880d-6762-4d57-881d-bc585318d500") : configmap "openstack-scripts" not found Nov 28 11:47:31 crc kubenswrapper[4862]: I1128 11:47:31.856358 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ps8s" event={"ID":"8631bc3d-1854-4bfa-b14a-e68dca7567e5","Type":"ContainerStarted","Data":"1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b"} Nov 28 11:47:32 crc kubenswrapper[4862]: E1128 11:47:32.333954 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:32 crc kubenswrapper[4862]: E1128 11:47:32.334064 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts podName:dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:47:40.334041458 +0000 UTC m=+1537.866555419 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts") pod "neutronc4cb-account-delete-9jrbn" (UID: "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f") : configmap "openstack-scripts" not found Nov 28 11:47:32 crc kubenswrapper[4862]: I1128 11:47:32.870037 4862 generic.go:334] "Generic (PLEG): container finished" podID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerID="1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b" exitCode=0 Nov 28 11:47:32 crc kubenswrapper[4862]: I1128 11:47:32.870115 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ps8s" event={"ID":"8631bc3d-1854-4bfa-b14a-e68dca7567e5","Type":"ContainerDied","Data":"1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b"} Nov 28 11:47:33 crc kubenswrapper[4862]: E1128 11:47:33.355165 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:33 crc kubenswrapper[4862]: E1128 11:47:33.355596 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts podName:880d930f-ea3a-4a7b-a437-5b164fbea09e nodeName:}" failed. No retries permitted until 2025-11-28 11:47:41.355576102 +0000 UTC m=+1538.888090023 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts") pod "cinder68e8-account-delete-rv4w4" (UID: "880d930f-ea3a-4a7b-a437-5b164fbea09e") : configmap "openstack-scripts" not found Nov 28 11:47:33 crc kubenswrapper[4862]: I1128 11:47:33.883021 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ps8s" event={"ID":"8631bc3d-1854-4bfa-b14a-e68dca7567e5","Type":"ContainerStarted","Data":"170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6"} Nov 28 11:47:34 crc kubenswrapper[4862]: E1128 11:47:34.914598 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:34 crc kubenswrapper[4862]: E1128 11:47:34.915224 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:34 crc kubenswrapper[4862]: E1128 11:47:34.915518 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:34 crc kubenswrapper[4862]: E1128 11:47:34.915563 4862 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" Nov 28 11:47:34 crc kubenswrapper[4862]: E1128 11:47:34.916661 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:34 crc kubenswrapper[4862]: E1128 11:47:34.918289 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:34 crc kubenswrapper[4862]: E1128 11:47:34.919687 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:34 crc kubenswrapper[4862]: E1128 11:47:34.919724 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.862380 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.893146 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5ps8s" podStartSLOduration=8.191161524 podStartE2EDuration="10.893130583s" podCreationTimestamp="2025-11-28 11:47:28 +0000 UTC" firstStartedPulling="2025-11-28 11:47:30.849828785 +0000 UTC m=+1528.382342706" lastFinishedPulling="2025-11-28 11:47:33.551797844 +0000 UTC m=+1531.084311765" observedRunningTime="2025-11-28 11:47:33.925367549 +0000 UTC m=+1531.457881470" watchObservedRunningTime="2025-11-28 11:47:38.893130583 +0000 UTC m=+1536.425644504" Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.930501 4862 generic.go:334] "Generic (PLEG): container finished" podID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerID="9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab" exitCode=0 Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.930542 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc889566f-77hs4" event={"ID":"a42b2a07-23e7-4281-b69b-e47635bcccf6","Type":"ContainerDied","Data":"9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab"} Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.930593 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dc889566f-77hs4" Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.930633 4862 scope.go:117] "RemoveContainer" containerID="76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f" Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.930618 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc889566f-77hs4" event={"ID":"a42b2a07-23e7-4281-b69b-e47635bcccf6","Type":"ContainerDied","Data":"7445e87e625eae3d947ef68ff946f6761065fa7f935e089b7a3405bc6fcb97b1"} Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.955038 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-combined-ca-bundle\") pod \"a42b2a07-23e7-4281-b69b-e47635bcccf6\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.955122 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-ovndb-tls-certs\") pod \"a42b2a07-23e7-4281-b69b-e47635bcccf6\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.955202 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-public-tls-certs\") pod \"a42b2a07-23e7-4281-b69b-e47635bcccf6\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.955240 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-httpd-config\") pod \"a42b2a07-23e7-4281-b69b-e47635bcccf6\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.955283 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-config\") pod \"a42b2a07-23e7-4281-b69b-e47635bcccf6\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.955311 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-internal-tls-certs\") pod \"a42b2a07-23e7-4281-b69b-e47635bcccf6\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.955995 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx27s\" (UniqueName: \"kubernetes.io/projected/a42b2a07-23e7-4281-b69b-e47635bcccf6-kube-api-access-jx27s\") pod \"a42b2a07-23e7-4281-b69b-e47635bcccf6\" (UID: \"a42b2a07-23e7-4281-b69b-e47635bcccf6\") " Nov 28 11:47:38 crc kubenswrapper[4862]: E1128 11:47:38.956414 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:38 crc kubenswrapper[4862]: E1128 11:47:38.956484 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts podName:409c7014-982c-4c60-b374-0ff46c398db4 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:54.956464841 +0000 UTC m=+1552.488978762 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts") pod "novaapi144c-account-delete-ls9c4" (UID: "409c7014-982c-4c60-b374-0ff46c398db4") : configmap "openstack-scripts" not found Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.960016 4862 scope.go:117] "RemoveContainer" containerID="9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab" Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.960953 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a42b2a07-23e7-4281-b69b-e47635bcccf6-kube-api-access-jx27s" (OuterVolumeSpecName: "kube-api-access-jx27s") pod "a42b2a07-23e7-4281-b69b-e47635bcccf6" (UID: "a42b2a07-23e7-4281-b69b-e47635bcccf6"). InnerVolumeSpecName "kube-api-access-jx27s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:38 crc kubenswrapper[4862]: I1128 11:47:38.972266 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a42b2a07-23e7-4281-b69b-e47635bcccf6" (UID: "a42b2a07-23e7-4281-b69b-e47635bcccf6"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.002726 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-config" (OuterVolumeSpecName: "config") pod "a42b2a07-23e7-4281-b69b-e47635bcccf6" (UID: "a42b2a07-23e7-4281-b69b-e47635bcccf6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.016369 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a42b2a07-23e7-4281-b69b-e47635bcccf6" (UID: "a42b2a07-23e7-4281-b69b-e47635bcccf6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.019458 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a42b2a07-23e7-4281-b69b-e47635bcccf6" (UID: "a42b2a07-23e7-4281-b69b-e47635bcccf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.039604 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a42b2a07-23e7-4281-b69b-e47635bcccf6" (UID: "a42b2a07-23e7-4281-b69b-e47635bcccf6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.042532 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a42b2a07-23e7-4281-b69b-e47635bcccf6" (UID: "a42b2a07-23e7-4281-b69b-e47635bcccf6"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.059522 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx27s\" (UniqueName: \"kubernetes.io/projected/a42b2a07-23e7-4281-b69b-e47635bcccf6-kube-api-access-jx27s\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.059576 4862 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.059594 4862 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.059651 4862 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.059650 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.059666 4862 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.059728 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts podName:ed3d880d-6762-4d57-881d-bc585318d500 nodeName:}" failed. No retries permitted until 2025-11-28 11:47:55.059708279 +0000 UTC m=+1552.592222200 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts") pod "barbican8d34-account-delete-bdd2b" (UID: "ed3d880d-6762-4d57-881d-bc585318d500") : configmap "openstack-scripts" not found Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.059756 4862 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-config\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.059768 4862 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42b2a07-23e7-4281-b69b-e47635bcccf6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.084111 4862 scope.go:117] "RemoveContainer" containerID="76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f" Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.084808 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f\": container with ID starting with 76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f not found: ID does not exist" containerID="76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.084861 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f"} err="failed to get container status \"76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f\": rpc error: code = NotFound desc = could not find container \"76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f\": container with ID starting with 76dc6bbcb60919c788ddb5f9c4519262a482f0ca5194f8c5f529a3657a00160f not found: ID does not exist" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.084892 4862 scope.go:117] "RemoveContainer" containerID="9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab" Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.085351 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab\": container with ID starting with 9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab not found: ID does not exist" containerID="9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.085385 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab"} err="failed to get container status \"9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab\": rpc error: code = NotFound desc = could not find container \"9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab\": container with ID starting with 9624f4ecacb0eb1dee874736c6d09a000c487478f6bea079281e02a1beaa21ab not found: ID does not exist" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.286167 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6dc889566f-77hs4"] Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.295454 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6dc889566f-77hs4"] Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.374216 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.375867 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:39 crc kubenswrapper[4862]: I1128 11:47:39.462213 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.915234 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.916298 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.916560 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.916600 4862 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.918109 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.920502 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.923024 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:39 crc kubenswrapper[4862]: E1128 11:47:39.923074 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" Nov 28 11:47:40 crc kubenswrapper[4862]: I1128 11:47:40.013730 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:40 crc kubenswrapper[4862]: I1128 11:47:40.091236 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5ps8s"] Nov 28 11:47:40 crc kubenswrapper[4862]: E1128 11:47:40.382964 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:40 crc kubenswrapper[4862]: E1128 11:47:40.383168 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts podName:dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:47:56.383118841 +0000 UTC m=+1553.915632802 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts") pod "neutronc4cb-account-delete-9jrbn" (UID: "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f") : configmap "openstack-scripts" not found Nov 28 11:47:40 crc kubenswrapper[4862]: I1128 11:47:40.850195 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" path="/var/lib/kubelet/pods/a42b2a07-23e7-4281-b69b-e47635bcccf6/volumes" Nov 28 11:47:41 crc kubenswrapper[4862]: E1128 11:47:41.410984 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:41 crc kubenswrapper[4862]: E1128 11:47:41.411425 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts podName:880d930f-ea3a-4a7b-a437-5b164fbea09e nodeName:}" failed. No retries permitted until 2025-11-28 11:47:57.411396358 +0000 UTC m=+1554.943910279 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts") pod "cinder68e8-account-delete-rv4w4" (UID: "880d930f-ea3a-4a7b-a437-5b164fbea09e") : configmap "openstack-scripts" not found Nov 28 11:47:41 crc kubenswrapper[4862]: I1128 11:47:41.971086 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5ps8s" podUID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerName="registry-server" containerID="cri-o://170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6" gracePeriod=2 Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.738993 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.744954 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrwb6\" (UniqueName: \"kubernetes.io/projected/8631bc3d-1854-4bfa-b14a-e68dca7567e5-kube-api-access-rrwb6\") pod \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.745013 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-utilities\") pod \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.745033 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-catalog-content\") pod \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\" (UID: \"8631bc3d-1854-4bfa-b14a-e68dca7567e5\") " Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.747954 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-utilities" (OuterVolumeSpecName: "utilities") pod "8631bc3d-1854-4bfa-b14a-e68dca7567e5" (UID: "8631bc3d-1854-4bfa-b14a-e68dca7567e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.762965 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8631bc3d-1854-4bfa-b14a-e68dca7567e5-kube-api-access-rrwb6" (OuterVolumeSpecName: "kube-api-access-rrwb6") pod "8631bc3d-1854-4bfa-b14a-e68dca7567e5" (UID: "8631bc3d-1854-4bfa-b14a-e68dca7567e5"). InnerVolumeSpecName "kube-api-access-rrwb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.820037 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8631bc3d-1854-4bfa-b14a-e68dca7567e5" (UID: "8631bc3d-1854-4bfa-b14a-e68dca7567e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.846494 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.846635 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631bc3d-1854-4bfa-b14a-e68dca7567e5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:43 crc kubenswrapper[4862]: I1128 11:47:43.846720 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrwb6\" (UniqueName: \"kubernetes.io/projected/8631bc3d-1854-4bfa-b14a-e68dca7567e5-kube-api-access-rrwb6\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.019199 4862 generic.go:334] "Generic (PLEG): container finished" podID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerID="170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6" exitCode=0 Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.019240 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ps8s" event={"ID":"8631bc3d-1854-4bfa-b14a-e68dca7567e5","Type":"ContainerDied","Data":"170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6"} Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.019269 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ps8s" event={"ID":"8631bc3d-1854-4bfa-b14a-e68dca7567e5","Type":"ContainerDied","Data":"21bcebc1a07581ebaf8ecfea5d759ecb32c5abd1d006d78834c9d9148e1f57ea"} Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.019285 4862 scope.go:117] "RemoveContainer" containerID="170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.019391 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5ps8s" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.044816 4862 scope.go:117] "RemoveContainer" containerID="1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.076671 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5ps8s"] Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.088466 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5ps8s"] Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.092045 4862 scope.go:117] "RemoveContainer" containerID="34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.114441 4862 scope.go:117] "RemoveContainer" containerID="170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6" Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.114857 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6\": container with ID starting with 170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6 not found: ID does not exist" containerID="170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.114899 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6"} err="failed to get container status \"170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6\": rpc error: code = NotFound desc = could not find container \"170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6\": container with ID starting with 170c4ce635f7b96c4df1ec00f260615985c59220f367f076942b9d5c18dbc5f6 not found: ID does not exist" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.114927 4862 scope.go:117] "RemoveContainer" containerID="1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b" Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.115198 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b\": container with ID starting with 1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b not found: ID does not exist" containerID="1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.115305 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b"} err="failed to get container status \"1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b\": rpc error: code = NotFound desc = could not find container \"1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b\": container with ID starting with 1c933980448471a2af388fecad7777596433af310ee2f15fe71a542d168b3e5b not found: ID does not exist" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.115396 4862 scope.go:117] "RemoveContainer" containerID="34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3" Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.115883 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3\": container with ID starting with 34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3 not found: ID does not exist" containerID="34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.115948 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3"} err="failed to get container status \"34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3\": rpc error: code = NotFound desc = could not find container \"34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3\": container with ID starting with 34185f86c99eb60b7585bcf10d668e4a5a8342ae1234425554f04086278109e3 not found: ID does not exist" Nov 28 11:47:44 crc kubenswrapper[4862]: I1128 11:47:44.856183 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" path="/var/lib/kubelet/pods/8631bc3d-1854-4bfa-b14a-e68dca7567e5/volumes" Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.914583 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.915322 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.915724 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.915810 4862 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.916487 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.919149 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.922606 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:44 crc kubenswrapper[4862]: E1128 11:47:44.922701 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" Nov 28 11:47:49 crc kubenswrapper[4862]: E1128 11:47:49.914327 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:49 crc kubenswrapper[4862]: E1128 11:47:49.915727 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:49 crc kubenswrapper[4862]: E1128 11:47:49.916315 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 11:47:49 crc kubenswrapper[4862]: E1128 11:47:49.916368 4862 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" Nov 28 11:47:49 crc kubenswrapper[4862]: E1128 11:47:49.916419 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:49 crc kubenswrapper[4862]: E1128 11:47:49.918427 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:49 crc kubenswrapper[4862]: E1128 11:47:49.920477 4862 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 11:47:49 crc kubenswrapper[4862]: E1128 11:47:49.920568 4862 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-hdvv6" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.105596 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hdvv6_ba7cc6b9-1443-45de-84f4-4ec3b92e699c/ovs-vswitchd/0.log" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.106914 4862 generic.go:334] "Generic (PLEG): container finished" podID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" exitCode=137 Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.106973 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hdvv6" event={"ID":"ba7cc6b9-1443-45de-84f4-4ec3b92e699c","Type":"ContainerDied","Data":"c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3"} Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.857937 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.876462 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9lp4\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-kube-api-access-l9lp4\") pod \"b56372b7-fc39-40cc-8e0a-e95f907db697\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.876560 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-lock\") pod \"b56372b7-fc39-40cc-8e0a-e95f907db697\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.876592 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"b56372b7-fc39-40cc-8e0a-e95f907db697\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.876618 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") pod \"b56372b7-fc39-40cc-8e0a-e95f907db697\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.876708 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-cache\") pod \"b56372b7-fc39-40cc-8e0a-e95f907db697\" (UID: \"b56372b7-fc39-40cc-8e0a-e95f907db697\") " Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.877485 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-lock" (OuterVolumeSpecName: "lock") pod "b56372b7-fc39-40cc-8e0a-e95f907db697" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.877625 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-cache" (OuterVolumeSpecName: "cache") pod "b56372b7-fc39-40cc-8e0a-e95f907db697" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.883284 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b56372b7-fc39-40cc-8e0a-e95f907db697" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.883339 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-kube-api-access-l9lp4" (OuterVolumeSpecName: "kube-api-access-l9lp4") pod "b56372b7-fc39-40cc-8e0a-e95f907db697" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697"). InnerVolumeSpecName "kube-api-access-l9lp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.884931 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "swift") pod "b56372b7-fc39-40cc-8e0a-e95f907db697" (UID: "b56372b7-fc39-40cc-8e0a-e95f907db697"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.975937 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hdvv6_ba7cc6b9-1443-45de-84f4-4ec3b92e699c/ovs-vswitchd/0.log" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.977339 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.978051 4862 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-cache\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.978085 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9lp4\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-kube-api-access-l9lp4\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.978136 4862 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b56372b7-fc39-40cc-8e0a-e95f907db697-lock\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.978176 4862 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 11:47:50 crc kubenswrapper[4862]: I1128 11:47:50.978189 4862 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b56372b7-fc39-40cc-8e0a-e95f907db697-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.007032 4862 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079186 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtpb9\" (UniqueName: \"kubernetes.io/projected/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-kube-api-access-jtpb9\") pod \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079242 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-log\") pod \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079349 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-run\") pod \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079440 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-etc-ovs\") pod \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079469 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-lib\") pod \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079431 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-log" (OuterVolumeSpecName: "var-log") pod "ba7cc6b9-1443-45de-84f4-4ec3b92e699c" (UID: "ba7cc6b9-1443-45de-84f4-4ec3b92e699c"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079466 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-run" (OuterVolumeSpecName: "var-run") pod "ba7cc6b9-1443-45de-84f4-4ec3b92e699c" (UID: "ba7cc6b9-1443-45de-84f4-4ec3b92e699c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079535 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-scripts\") pod \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\" (UID: \"ba7cc6b9-1443-45de-84f4-4ec3b92e699c\") " Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079531 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "ba7cc6b9-1443-45de-84f4-4ec3b92e699c" (UID: "ba7cc6b9-1443-45de-84f4-4ec3b92e699c"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079584 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-lib" (OuterVolumeSpecName: "var-lib") pod "ba7cc6b9-1443-45de-84f4-4ec3b92e699c" (UID: "ba7cc6b9-1443-45de-84f4-4ec3b92e699c"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.079987 4862 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.080011 4862 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-etc-ovs\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.080026 4862 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-lib\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.080041 4862 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.080056 4862 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-var-log\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.080738 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-scripts" (OuterVolumeSpecName: "scripts") pod "ba7cc6b9-1443-45de-84f4-4ec3b92e699c" (UID: "ba7cc6b9-1443-45de-84f4-4ec3b92e699c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.083892 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-kube-api-access-jtpb9" (OuterVolumeSpecName: "kube-api-access-jtpb9") pod "ba7cc6b9-1443-45de-84f4-4ec3b92e699c" (UID: "ba7cc6b9-1443-45de-84f4-4ec3b92e699c"). InnerVolumeSpecName "kube-api-access-jtpb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.119154 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hdvv6_ba7cc6b9-1443-45de-84f4-4ec3b92e699c/ovs-vswitchd/0.log" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.120223 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-hdvv6" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.120227 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hdvv6" event={"ID":"ba7cc6b9-1443-45de-84f4-4ec3b92e699c","Type":"ContainerDied","Data":"46f4a765ad8066346562b5840ee471cf98f1740cbd169e69b17951f103ae325e"} Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.120281 4862 scope.go:117] "RemoveContainer" containerID="c52618d82244a9be826a58c6fa59aad3cee4577989cb9dbcbef29dee57c829f3" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.129611 4862 generic.go:334] "Generic (PLEG): container finished" podID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerID="c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255" exitCode=137 Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.129656 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255"} Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.129685 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b56372b7-fc39-40cc-8e0a-e95f907db697","Type":"ContainerDied","Data":"7afc02dbfce6529a27ddb4ccc07fef4871e0b156b293bff22ddb94283f95f8c4"} Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.129813 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.172268 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-hdvv6"] Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.178895 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-hdvv6"] Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.184911 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtpb9\" (UniqueName: \"kubernetes.io/projected/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-kube-api-access-jtpb9\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.184952 4862 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba7cc6b9-1443-45de-84f4-4ec3b92e699c-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.198824 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.204119 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.320782 4862 scope.go:117] "RemoveContainer" containerID="49dee8d3435b08e294bdc41cb68a46945a0e2f282a088a0b9deab519b2b8a5d8" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.351018 4862 scope.go:117] "RemoveContainer" containerID="b26a7efa52fe59cc26535571cd7c3b5c387aba3ffa9eb88168ffcfe54e853cf0" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.377326 4862 scope.go:117] "RemoveContainer" containerID="c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.427772 4862 scope.go:117] "RemoveContainer" containerID="775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.462858 4862 scope.go:117] "RemoveContainer" containerID="0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.493369 4862 scope.go:117] "RemoveContainer" containerID="84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.532656 4862 scope.go:117] "RemoveContainer" containerID="2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.569264 4862 scope.go:117] "RemoveContainer" containerID="891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.605530 4862 scope.go:117] "RemoveContainer" containerID="4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.635867 4862 scope.go:117] "RemoveContainer" containerID="f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.662165 4862 scope.go:117] "RemoveContainer" containerID="df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.686445 4862 scope.go:117] "RemoveContainer" containerID="d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.714472 4862 scope.go:117] "RemoveContainer" containerID="563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.739141 4862 scope.go:117] "RemoveContainer" containerID="53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.765020 4862 scope.go:117] "RemoveContainer" containerID="aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.793800 4862 scope.go:117] "RemoveContainer" containerID="b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.822016 4862 scope.go:117] "RemoveContainer" containerID="540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.850607 4862 scope.go:117] "RemoveContainer" containerID="c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.851164 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255\": container with ID starting with c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255 not found: ID does not exist" containerID="c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.851232 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255"} err="failed to get container status \"c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255\": rpc error: code = NotFound desc = could not find container \"c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255\": container with ID starting with c3bb49140178e3a5d81c14c9d091d307cfffa07b38f3d24694e284d46170f255 not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.851276 4862 scope.go:117] "RemoveContainer" containerID="775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.851874 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6\": container with ID starting with 775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6 not found: ID does not exist" containerID="775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.851936 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6"} err="failed to get container status \"775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6\": rpc error: code = NotFound desc = could not find container \"775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6\": container with ID starting with 775255a51961b0f02ed961d20d122d0db97f5df6135db2171d0c756588500eb6 not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.851964 4862 scope.go:117] "RemoveContainer" containerID="0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.852677 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3\": container with ID starting with 0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3 not found: ID does not exist" containerID="0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.852753 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3"} err="failed to get container status \"0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3\": rpc error: code = NotFound desc = could not find container \"0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3\": container with ID starting with 0fca1a96d1b3f6b2bb8b6f9c8c50e5b83bfa48126f37056d9b6f4946cce8f2c3 not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.852799 4862 scope.go:117] "RemoveContainer" containerID="84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.853308 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d\": container with ID starting with 84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d not found: ID does not exist" containerID="84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.853354 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d"} err="failed to get container status \"84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d\": rpc error: code = NotFound desc = could not find container \"84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d\": container with ID starting with 84bf73d55b5bac95ed3939aa9199dd62e11aac1107043c1b8352a9a023dbbc6d not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.853385 4862 scope.go:117] "RemoveContainer" containerID="2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.854040 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97\": container with ID starting with 2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97 not found: ID does not exist" containerID="2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.854082 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97"} err="failed to get container status \"2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97\": rpc error: code = NotFound desc = could not find container \"2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97\": container with ID starting with 2c06f9c8eb6ef73cf0eadb0bc563c763b837be3bed06e74f7e9ab8cb8fe06b97 not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.854145 4862 scope.go:117] "RemoveContainer" containerID="891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.854686 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c\": container with ID starting with 891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c not found: ID does not exist" containerID="891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.854752 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c"} err="failed to get container status \"891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c\": rpc error: code = NotFound desc = could not find container \"891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c\": container with ID starting with 891a8eda8d103a96225935513f6f5041970270a8e03fb11054acb85220d0d16c not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.854791 4862 scope.go:117] "RemoveContainer" containerID="4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.855257 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02\": container with ID starting with 4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02 not found: ID does not exist" containerID="4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.855308 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02"} err="failed to get container status \"4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02\": rpc error: code = NotFound desc = could not find container \"4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02\": container with ID starting with 4906e5599c88b1e18ed8338687477530be77c548a751f7933abd08cb3016fa02 not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.855339 4862 scope.go:117] "RemoveContainer" containerID="f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.855843 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726\": container with ID starting with f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726 not found: ID does not exist" containerID="f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.855890 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726"} err="failed to get container status \"f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726\": rpc error: code = NotFound desc = could not find container \"f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726\": container with ID starting with f8b53d385fd45518b19787ee123d45d457d4a6af03c386f37cf879d171ef1726 not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.855922 4862 scope.go:117] "RemoveContainer" containerID="df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.856382 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a\": container with ID starting with df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a not found: ID does not exist" containerID="df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.856429 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a"} err="failed to get container status \"df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a\": rpc error: code = NotFound desc = could not find container \"df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a\": container with ID starting with df2272b33e0d6132c1643b9ac9ffadf3b0b8be9bdc1f658b9f5710233ebe896a not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.856457 4862 scope.go:117] "RemoveContainer" containerID="d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.856978 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61\": container with ID starting with d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61 not found: ID does not exist" containerID="d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.857032 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61"} err="failed to get container status \"d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61\": rpc error: code = NotFound desc = could not find container \"d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61\": container with ID starting with d7c6ff96ce648b96f78da8a411101ea12a7d5e0630dd11a5f934aef5148a8a61 not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.857066 4862 scope.go:117] "RemoveContainer" containerID="563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.857620 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235\": container with ID starting with 563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235 not found: ID does not exist" containerID="563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.857666 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235"} err="failed to get container status \"563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235\": rpc error: code = NotFound desc = could not find container \"563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235\": container with ID starting with 563721ef848cf8da13f7df2d6dd7026327a81642295289d5450ddfedc4096235 not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.857700 4862 scope.go:117] "RemoveContainer" containerID="53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.858399 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a\": container with ID starting with 53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a not found: ID does not exist" containerID="53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.858467 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a"} err="failed to get container status \"53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a\": rpc error: code = NotFound desc = could not find container \"53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a\": container with ID starting with 53528c494933ef36d2432a53820d943f3f49f61aa800412ae4bb1760e630888a not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.858506 4862 scope.go:117] "RemoveContainer" containerID="aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.859265 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3\": container with ID starting with aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3 not found: ID does not exist" containerID="aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.859311 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3"} err="failed to get container status \"aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3\": rpc error: code = NotFound desc = could not find container \"aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3\": container with ID starting with aeb39a3a3fcbad07a61e90493266e3b7c74d2de57977736cd9bb4984e3b449e3 not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.859340 4862 scope.go:117] "RemoveContainer" containerID="b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.859776 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc\": container with ID starting with b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc not found: ID does not exist" containerID="b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.859827 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc"} err="failed to get container status \"b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc\": rpc error: code = NotFound desc = could not find container \"b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc\": container with ID starting with b8749dce6ded7d406c66d35baf9cbc69fe736441994cdeeebb5a50d7535e04dc not found: ID does not exist" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.859856 4862 scope.go:117] "RemoveContainer" containerID="540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e" Nov 28 11:47:51 crc kubenswrapper[4862]: E1128 11:47:51.860348 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e\": container with ID starting with 540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e not found: ID does not exist" containerID="540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e" Nov 28 11:47:51 crc kubenswrapper[4862]: I1128 11:47:51.860389 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e"} err="failed to get container status \"540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e\": rpc error: code = NotFound desc = could not find container \"540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e\": container with ID starting with 540e155ac73894919ad90baadf49fcb5151b4f795d45762127dc2c64d3e38b6e not found: ID does not exist" Nov 28 11:47:52 crc kubenswrapper[4862]: I1128 11:47:52.856780 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" path="/var/lib/kubelet/pods/b56372b7-fc39-40cc-8e0a-e95f907db697/volumes" Nov 28 11:47:52 crc kubenswrapper[4862]: I1128 11:47:52.860305 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" path="/var/lib/kubelet/pods/ba7cc6b9-1443-45de-84f4-4ec3b92e699c/volumes" Nov 28 11:47:53 crc kubenswrapper[4862]: I1128 11:47:53.703410 4862 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod23891b0e-3fc7-4d2f-9abf-96b57b09ca01"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod23891b0e-3fc7-4d2f-9abf-96b57b09ca01] : Timed out while waiting for systemd to remove kubepods-besteffort-pod23891b0e_3fc7_4d2f_9abf_96b57b09ca01.slice" Nov 28 11:47:53 crc kubenswrapper[4862]: E1128 11:47:53.703905 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod23891b0e-3fc7-4d2f-9abf-96b57b09ca01] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod23891b0e-3fc7-4d2f-9abf-96b57b09ca01] : Timed out while waiting for systemd to remove kubepods-besteffort-pod23891b0e_3fc7_4d2f_9abf_96b57b09ca01.slice" pod="openstack/openstackclient" podUID="23891b0e-3fc7-4d2f-9abf-96b57b09ca01" Nov 28 11:47:54 crc kubenswrapper[4862]: I1128 11:47:54.167291 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 11:47:55 crc kubenswrapper[4862]: E1128 11:47:55.055755 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:55 crc kubenswrapper[4862]: E1128 11:47:55.055887 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts podName:409c7014-982c-4c60-b374-0ff46c398db4 nodeName:}" failed. No retries permitted until 2025-11-28 11:48:27.055852724 +0000 UTC m=+1584.588366685 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts") pod "novaapi144c-account-delete-ls9c4" (UID: "409c7014-982c-4c60-b374-0ff46c398db4") : configmap "openstack-scripts" not found Nov 28 11:47:55 crc kubenswrapper[4862]: E1128 11:47:55.157529 4862 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 11:47:55 crc kubenswrapper[4862]: E1128 11:47:55.157691 4862 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts podName:ed3d880d-6762-4d57-881d-bc585318d500 nodeName:}" failed. No retries permitted until 2025-11-28 11:48:27.15764546 +0000 UTC m=+1584.690159431 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts") pod "barbican8d34-account-delete-bdd2b" (UID: "ed3d880d-6762-4d57-881d-bc585318d500") : configmap "openstack-scripts" not found Nov 28 11:47:55 crc kubenswrapper[4862]: I1128 11:47:55.985634 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.031767 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.038804 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.044870 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.069749 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qqsr\" (UniqueName: \"kubernetes.io/projected/409c7014-982c-4c60-b374-0ff46c398db4-kube-api-access-2qqsr\") pod \"409c7014-982c-4c60-b374-0ff46c398db4\" (UID: \"409c7014-982c-4c60-b374-0ff46c398db4\") " Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.069819 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts\") pod \"ed3d880d-6762-4d57-881d-bc585318d500\" (UID: \"ed3d880d-6762-4d57-881d-bc585318d500\") " Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.069852 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71c0d0c9-ed19-41a0-a3d1-92088179bf04-operator-scripts\") pod \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\" (UID: \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\") " Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.069910 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts\") pod \"409c7014-982c-4c60-b374-0ff46c398db4\" (UID: \"409c7014-982c-4c60-b374-0ff46c398db4\") " Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.069928 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxn4v\" (UniqueName: \"kubernetes.io/projected/ed3d880d-6762-4d57-881d-bc585318d500-kube-api-access-jxn4v\") pod \"ed3d880d-6762-4d57-881d-bc585318d500\" (UID: \"ed3d880d-6762-4d57-881d-bc585318d500\") " Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.069953 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j44b\" (UniqueName: \"kubernetes.io/projected/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-kube-api-access-8j44b\") pod \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\" (UID: \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\") " Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.069997 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zqg7\" (UniqueName: \"kubernetes.io/projected/71c0d0c9-ed19-41a0-a3d1-92088179bf04-kube-api-access-4zqg7\") pod \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\" (UID: \"71c0d0c9-ed19-41a0-a3d1-92088179bf04\") " Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.070043 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts\") pod \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\" (UID: \"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f\") " Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.070769 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "409c7014-982c-4c60-b374-0ff46c398db4" (UID: "409c7014-982c-4c60-b374-0ff46c398db4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.070813 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f" (UID: "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.070862 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71c0d0c9-ed19-41a0-a3d1-92088179bf04-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "71c0d0c9-ed19-41a0-a3d1-92088179bf04" (UID: "71c0d0c9-ed19-41a0-a3d1-92088179bf04"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.071193 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed3d880d-6762-4d57-881d-bc585318d500" (UID: "ed3d880d-6762-4d57-881d-bc585318d500"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.079019 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-kube-api-access-8j44b" (OuterVolumeSpecName: "kube-api-access-8j44b") pod "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f" (UID: "dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f"). InnerVolumeSpecName "kube-api-access-8j44b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.079179 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed3d880d-6762-4d57-881d-bc585318d500-kube-api-access-jxn4v" (OuterVolumeSpecName: "kube-api-access-jxn4v") pod "ed3d880d-6762-4d57-881d-bc585318d500" (UID: "ed3d880d-6762-4d57-881d-bc585318d500"). InnerVolumeSpecName "kube-api-access-jxn4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.079582 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/409c7014-982c-4c60-b374-0ff46c398db4-kube-api-access-2qqsr" (OuterVolumeSpecName: "kube-api-access-2qqsr") pod "409c7014-982c-4c60-b374-0ff46c398db4" (UID: "409c7014-982c-4c60-b374-0ff46c398db4"). InnerVolumeSpecName "kube-api-access-2qqsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.079967 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c0d0c9-ed19-41a0-a3d1-92088179bf04-kube-api-access-4zqg7" (OuterVolumeSpecName: "kube-api-access-4zqg7") pod "71c0d0c9-ed19-41a0-a3d1-92088179bf04" (UID: "71c0d0c9-ed19-41a0-a3d1-92088179bf04"). InnerVolumeSpecName "kube-api-access-4zqg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.171970 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/409c7014-982c-4c60-b374-0ff46c398db4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.172005 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxn4v\" (UniqueName: \"kubernetes.io/projected/ed3d880d-6762-4d57-881d-bc585318d500-kube-api-access-jxn4v\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.172015 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j44b\" (UniqueName: \"kubernetes.io/projected/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-kube-api-access-8j44b\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.172025 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zqg7\" (UniqueName: \"kubernetes.io/projected/71c0d0c9-ed19-41a0-a3d1-92088179bf04-kube-api-access-4zqg7\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.172035 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.172046 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qqsr\" (UniqueName: \"kubernetes.io/projected/409c7014-982c-4c60-b374-0ff46c398db4-kube-api-access-2qqsr\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.172056 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3d880d-6762-4d57-881d-bc585318d500-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.172064 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71c0d0c9-ed19-41a0-a3d1-92088179bf04-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.190174 4862 generic.go:334] "Generic (PLEG): container finished" podID="dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f" containerID="9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b" exitCode=137 Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.190210 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronc4cb-account-delete-9jrbn" event={"ID":"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f","Type":"ContainerDied","Data":"9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b"} Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.190269 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronc4cb-account-delete-9jrbn" event={"ID":"dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f","Type":"ContainerDied","Data":"ef6743975efa11f2e33cf5ac1a219f311371830c64fb99a93e655444f7507f5c"} Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.190294 4862 scope.go:117] "RemoveContainer" containerID="9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.190201 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronc4cb-account-delete-9jrbn" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.194084 4862 generic.go:334] "Generic (PLEG): container finished" podID="409c7014-982c-4c60-b374-0ff46c398db4" containerID="731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a" exitCode=137 Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.194142 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi144c-account-delete-ls9c4" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.194190 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi144c-account-delete-ls9c4" event={"ID":"409c7014-982c-4c60-b374-0ff46c398db4","Type":"ContainerDied","Data":"731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a"} Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.194219 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi144c-account-delete-ls9c4" event={"ID":"409c7014-982c-4c60-b374-0ff46c398db4","Type":"ContainerDied","Data":"13721b6da26a5c16054dc852bfdf1a7f3824a0349e3c435a25f2b19f24b7e29f"} Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.197278 4862 generic.go:334] "Generic (PLEG): container finished" podID="ed3d880d-6762-4d57-881d-bc585318d500" containerID="3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698" exitCode=137 Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.197370 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8d34-account-delete-bdd2b" event={"ID":"ed3d880d-6762-4d57-881d-bc585318d500","Type":"ContainerDied","Data":"3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698"} Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.197443 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8d34-account-delete-bdd2b" event={"ID":"ed3d880d-6762-4d57-881d-bc585318d500","Type":"ContainerDied","Data":"b831312cf6684723dbfa97036d2f9334dda17fd663f427687ccd6143f9604a65"} Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.197497 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8d34-account-delete-bdd2b" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.199996 4862 generic.go:334] "Generic (PLEG): container finished" podID="71c0d0c9-ed19-41a0-a3d1-92088179bf04" containerID="83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772" exitCode=137 Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.200025 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance056c-account-delete-cn6wm" event={"ID":"71c0d0c9-ed19-41a0-a3d1-92088179bf04","Type":"ContainerDied","Data":"83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772"} Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.200042 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance056c-account-delete-cn6wm" event={"ID":"71c0d0c9-ed19-41a0-a3d1-92088179bf04","Type":"ContainerDied","Data":"ecde36974a2cbfb5129ab83ac86c44e5da8699f995d1ff184cf6a50991b0ea4e"} Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.200099 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance056c-account-delete-cn6wm" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.219992 4862 scope.go:117] "RemoveContainer" containerID="9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b" Nov 28 11:47:56 crc kubenswrapper[4862]: E1128 11:47:56.220404 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b\": container with ID starting with 9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b not found: ID does not exist" containerID="9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.220480 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b"} err="failed to get container status \"9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b\": rpc error: code = NotFound desc = could not find container \"9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b\": container with ID starting with 9fc01c8ad27a83095634cffaa2bcf62ed25f593aea6c896c560614ce38e8982b not found: ID does not exist" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.220530 4862 scope.go:117] "RemoveContainer" containerID="731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.251951 4862 scope.go:117] "RemoveContainer" containerID="731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a" Nov 28 11:47:56 crc kubenswrapper[4862]: E1128 11:47:56.267104 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a\": container with ID starting with 731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a not found: ID does not exist" containerID="731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.267514 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a"} err="failed to get container status \"731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a\": rpc error: code = NotFound desc = could not find container \"731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a\": container with ID starting with 731f84bd09c5272820ce56e4ae323eb709f89fb88fd11d8d9ae81b89d44f0b0a not found: ID does not exist" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.267608 4862 scope.go:117] "RemoveContainer" containerID="3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.268314 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi144c-account-delete-ls9c4"] Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.284587 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi144c-account-delete-ls9c4"] Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.294175 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8d34-account-delete-bdd2b"] Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.304655 4862 scope.go:117] "RemoveContainer" containerID="3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.305296 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican8d34-account-delete-bdd2b"] Nov 28 11:47:56 crc kubenswrapper[4862]: E1128 11:47:56.305311 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698\": container with ID starting with 3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698 not found: ID does not exist" containerID="3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.305345 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698"} err="failed to get container status \"3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698\": rpc error: code = NotFound desc = could not find container \"3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698\": container with ID starting with 3bdce3baea1b55dab170cf8fadb7e4933dab0c57f5623bebaf5bfd6be66e7698 not found: ID does not exist" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.305373 4862 scope.go:117] "RemoveContainer" containerID="83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.310693 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronc4cb-account-delete-9jrbn"] Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.315588 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutronc4cb-account-delete-9jrbn"] Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.320340 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance056c-account-delete-cn6wm"] Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.323289 4862 scope.go:117] "RemoveContainer" containerID="83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772" Nov 28 11:47:56 crc kubenswrapper[4862]: E1128 11:47:56.323663 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772\": container with ID starting with 83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772 not found: ID does not exist" containerID="83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.323701 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772"} err="failed to get container status \"83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772\": rpc error: code = NotFound desc = could not find container \"83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772\": container with ID starting with 83b6f9851ef9111e220b5350a2895ba0bbcf9b63b2308a84118653a369c72772 not found: ID does not exist" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.325794 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance056c-account-delete-cn6wm"] Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.858115 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="409c7014-982c-4c60-b374-0ff46c398db4" path="/var/lib/kubelet/pods/409c7014-982c-4c60-b374-0ff46c398db4/volumes" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.859141 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71c0d0c9-ed19-41a0-a3d1-92088179bf04" path="/var/lib/kubelet/pods/71c0d0c9-ed19-41a0-a3d1-92088179bf04/volumes" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.859648 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f" path="/var/lib/kubelet/pods/dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f/volumes" Nov 28 11:47:56 crc kubenswrapper[4862]: I1128 11:47:56.860108 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed3d880d-6762-4d57-881d-bc585318d500" path="/var/lib/kubelet/pods/ed3d880d-6762-4d57-881d-bc585318d500/volumes" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.201324 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.217357 4862 generic.go:334] "Generic (PLEG): container finished" podID="880d930f-ea3a-4a7b-a437-5b164fbea09e" containerID="4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790" exitCode=137 Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.217420 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder68e8-account-delete-rv4w4" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.217451 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder68e8-account-delete-rv4w4" event={"ID":"880d930f-ea3a-4a7b-a437-5b164fbea09e","Type":"ContainerDied","Data":"4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790"} Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.217518 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder68e8-account-delete-rv4w4" event={"ID":"880d930f-ea3a-4a7b-a437-5b164fbea09e","Type":"ContainerDied","Data":"3c0d559d544a271543807e2e3c76be510d522b2484c9c8682f300b0002ab68bf"} Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.217548 4862 scope.go:117] "RemoveContainer" containerID="4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.259857 4862 scope.go:117] "RemoveContainer" containerID="4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790" Nov 28 11:47:57 crc kubenswrapper[4862]: E1128 11:47:57.260625 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790\": container with ID starting with 4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790 not found: ID does not exist" containerID="4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.260687 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790"} err="failed to get container status \"4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790\": rpc error: code = NotFound desc = could not find container \"4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790\": container with ID starting with 4139629731fdc17524bf2b826c3db55eea185d6d6d02e311ee40e388dbd5a790 not found: ID does not exist" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.290523 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwb7s\" (UniqueName: \"kubernetes.io/projected/880d930f-ea3a-4a7b-a437-5b164fbea09e-kube-api-access-fwb7s\") pod \"880d930f-ea3a-4a7b-a437-5b164fbea09e\" (UID: \"880d930f-ea3a-4a7b-a437-5b164fbea09e\") " Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.290730 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts\") pod \"880d930f-ea3a-4a7b-a437-5b164fbea09e\" (UID: \"880d930f-ea3a-4a7b-a437-5b164fbea09e\") " Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.291464 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "880d930f-ea3a-4a7b-a437-5b164fbea09e" (UID: "880d930f-ea3a-4a7b-a437-5b164fbea09e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.296387 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880d930f-ea3a-4a7b-a437-5b164fbea09e-kube-api-access-fwb7s" (OuterVolumeSpecName: "kube-api-access-fwb7s") pod "880d930f-ea3a-4a7b-a437-5b164fbea09e" (UID: "880d930f-ea3a-4a7b-a437-5b164fbea09e"). InnerVolumeSpecName "kube-api-access-fwb7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.392644 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwb7s\" (UniqueName: \"kubernetes.io/projected/880d930f-ea3a-4a7b-a437-5b164fbea09e-kube-api-access-fwb7s\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.392686 4862 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880d930f-ea3a-4a7b-a437-5b164fbea09e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.555252 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder68e8-account-delete-rv4w4"] Nov 28 11:47:57 crc kubenswrapper[4862]: I1128 11:47:57.565144 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder68e8-account-delete-rv4w4"] Nov 28 11:47:58 crc kubenswrapper[4862]: I1128 11:47:58.855157 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="880d930f-ea3a-4a7b-a437-5b164fbea09e" path="/var/lib/kubelet/pods/880d930f-ea3a-4a7b-a437-5b164fbea09e/volumes" Nov 28 11:48:08 crc kubenswrapper[4862]: I1128 11:48:08.291974 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:48:08 crc kubenswrapper[4862]: I1128 11:48:08.292847 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:48:16 crc kubenswrapper[4862]: I1128 11:48:16.488876 4862 scope.go:117] "RemoveContainer" containerID="9039d4893d7c154aacf5e3d7aa83396bed1a0000865ee2be9f2d19e6c72246c2" Nov 28 11:48:16 crc kubenswrapper[4862]: I1128 11:48:16.550577 4862 scope.go:117] "RemoveContainer" containerID="4f561afbdd4fb1259076c496dea33b5413f902ebee541fb48eb7e1aa8fa45027" Nov 28 11:48:38 crc kubenswrapper[4862]: I1128 11:48:38.291825 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:48:38 crc kubenswrapper[4862]: I1128 11:48:38.292512 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:49:08 crc kubenswrapper[4862]: I1128 11:49:08.292329 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:49:08 crc kubenswrapper[4862]: I1128 11:49:08.293136 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:49:08 crc kubenswrapper[4862]: I1128 11:49:08.293218 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:49:08 crc kubenswrapper[4862]: I1128 11:49:08.294130 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776"} pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:49:08 crc kubenswrapper[4862]: I1128 11:49:08.294249 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" containerID="cri-o://990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" gracePeriod=600 Nov 28 11:49:08 crc kubenswrapper[4862]: E1128 11:49:08.429760 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:49:09 crc kubenswrapper[4862]: I1128 11:49:09.086657 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6845439-1a4c-4f28-9ba4-29149d545995" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" exitCode=0 Nov 28 11:49:09 crc kubenswrapper[4862]: I1128 11:49:09.086822 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerDied","Data":"990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776"} Nov 28 11:49:09 crc kubenswrapper[4862]: I1128 11:49:09.086892 4862 scope.go:117] "RemoveContainer" containerID="9f1769f4e9e7625f7b202c55e00f4690f331fdd2c1054c72cfd81ec88107ff67" Nov 28 11:49:09 crc kubenswrapper[4862]: I1128 11:49:09.089551 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:49:09 crc kubenswrapper[4862]: E1128 11:49:09.090784 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.394820 4862 scope.go:117] "RemoveContainer" containerID="8796e3b178d354b282dda262ad4d4df0cd3fe71afbe530ecd93777cdc96c0188" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.457466 4862 scope.go:117] "RemoveContainer" containerID="9c34d5f4b793116c2cc68fd196f337aca1ef159640d7e85b2d3fbe1065f54a98" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.488269 4862 scope.go:117] "RemoveContainer" containerID="1b4d99bd2494049894725880aae78e315feeedf9ec2f86b9e684dcb2edada6c1" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.528113 4862 scope.go:117] "RemoveContainer" containerID="b6aaf41414c9fd41c763d2add37d020134b5bae524760cad7b8d323eb5a18a35" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.574337 4862 scope.go:117] "RemoveContainer" containerID="ac31c3f95462a4c5062acc4cdb0ced218a95b344159616b29747888ef3fb38b8" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.612070 4862 scope.go:117] "RemoveContainer" containerID="11f076f97304361292994e05f8355c71c3532228c8bc7325739454e7865993d0" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.643382 4862 scope.go:117] "RemoveContainer" containerID="96915e9686bbfe96431db94d018ab010ba0fdc714bbd5170e37a594dd24d598b" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.663909 4862 scope.go:117] "RemoveContainer" containerID="d98f2a5cfade1b9e22b1a5356a69fd726d97f61e64058fa2b2bf4d0f31c331b9" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.682833 4862 scope.go:117] "RemoveContainer" containerID="12cc858922f84aa30fabdabce429fbe3a12d1c5905d381e7d6b7a6811417f1a1" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.705263 4862 scope.go:117] "RemoveContainer" containerID="00badb80af0f696931c741191dda1d4b64ccf8e89732147e270eb8afbaa447c3" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.742399 4862 scope.go:117] "RemoveContainer" containerID="0f28ad331cb738df8055d1c295bfeb6493d33b47ddd7640edef5f7af8a83c564" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.763582 4862 scope.go:117] "RemoveContainer" containerID="aedd0f47afe828a0180407f4e6ad432542ca5f49e5f4367abad5584e65aa27c7" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.786371 4862 scope.go:117] "RemoveContainer" containerID="f55bf280b2e3933ff62cacb88d6bf607a6df0aff1ec46aa2f1e8a94ca25b0425" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.810918 4862 scope.go:117] "RemoveContainer" containerID="97c0b76d1ae2e41b6a809bf0f557b15892c29ab5beabbc51aa8dc791239b70cb" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.829899 4862 scope.go:117] "RemoveContainer" containerID="fb97e0ded23986063a8aee20ab8624987a936bd554e2e3bf8bd543c73eed9c7a" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.848515 4862 scope.go:117] "RemoveContainer" containerID="fd35b665be415af27b736bab6a7dda350802c5432b416ed0360895ab52239829" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.882920 4862 scope.go:117] "RemoveContainer" containerID="ce8880d10b828a26bff152b7cdfba1672915a6ed5d72b43cd17a128c1264bfc5" Nov 28 11:49:17 crc kubenswrapper[4862]: I1128 11:49:17.918072 4862 scope.go:117] "RemoveContainer" containerID="e461f0dccb2ed908a6c4512b4a8acd9a84d85b485f10f9996cba1d1617b20a5d" Nov 28 11:49:21 crc kubenswrapper[4862]: I1128 11:49:21.839327 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:49:21 crc kubenswrapper[4862]: E1128 11:49:21.840279 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:49:32 crc kubenswrapper[4862]: I1128 11:49:32.844131 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:49:32 crc kubenswrapper[4862]: E1128 11:49:32.845072 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:49:44 crc kubenswrapper[4862]: I1128 11:49:44.839057 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:49:44 crc kubenswrapper[4862]: E1128 11:49:44.840331 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:49:58 crc kubenswrapper[4862]: I1128 11:49:58.839257 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:49:58 crc kubenswrapper[4862]: E1128 11:49:58.840288 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:50:10 crc kubenswrapper[4862]: I1128 11:50:10.838833 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:50:10 crc kubenswrapper[4862]: E1128 11:50:10.839896 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:50:18 crc kubenswrapper[4862]: I1128 11:50:18.300042 4862 scope.go:117] "RemoveContainer" containerID="cf25724267080641f4ff63f79418c992c069ccd67889b4a80b2782cf3e9a2d04" Nov 28 11:50:18 crc kubenswrapper[4862]: I1128 11:50:18.367805 4862 scope.go:117] "RemoveContainer" containerID="41c0bf7c013242ca225a1314624c3dc004d7184562ad1f7c5aa46b4bca21a2ae" Nov 28 11:50:18 crc kubenswrapper[4862]: I1128 11:50:18.427057 4862 scope.go:117] "RemoveContainer" containerID="24e2b7bdd8bb596c772aeefbf5a89bcacb1c66292c794bd6e2daecfa357f1740" Nov 28 11:50:18 crc kubenswrapper[4862]: I1128 11:50:18.467401 4862 scope.go:117] "RemoveContainer" containerID="0e41765288ac92ec73650d4eec94aa7bf446375875b6ddb400735a348f5b232a" Nov 28 11:50:18 crc kubenswrapper[4862]: I1128 11:50:18.515957 4862 scope.go:117] "RemoveContainer" containerID="57a230e1e4593a57f1af78d10623daad185b54c29ca745c19aa3ecc84aec7fbe" Nov 28 11:50:18 crc kubenswrapper[4862]: I1128 11:50:18.564950 4862 scope.go:117] "RemoveContainer" containerID="7e861fa3db125e3ec49410b4062ef8c9f7b09f017f185673a8589264da58f5ce" Nov 28 11:50:22 crc kubenswrapper[4862]: I1128 11:50:22.847207 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:50:22 crc kubenswrapper[4862]: E1128 11:50:22.848316 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:50:35 crc kubenswrapper[4862]: I1128 11:50:35.839676 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:50:35 crc kubenswrapper[4862]: E1128 11:50:35.840953 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:50:47 crc kubenswrapper[4862]: I1128 11:50:47.838009 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:50:47 crc kubenswrapper[4862]: E1128 11:50:47.838606 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:51:02 crc kubenswrapper[4862]: I1128 11:51:02.846791 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:51:02 crc kubenswrapper[4862]: E1128 11:51:02.848026 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:51:16 crc kubenswrapper[4862]: I1128 11:51:16.839217 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:51:16 crc kubenswrapper[4862]: E1128 11:51:16.840373 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.676350 4862 scope.go:117] "RemoveContainer" containerID="44535d7cc32f4331f5fc1f21419dfd1a2f555c351ab9752b1439e4c14dca91c9" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.706530 4862 scope.go:117] "RemoveContainer" containerID="b66e34c72b94210fa8e1b94676bc9d86fdb52829035fa8a84ba5bbf0ebf630c5" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.727141 4862 scope.go:117] "RemoveContainer" containerID="846a82b096c91fa279b8a2c68e6b0f950e08a8bae9936e49194d3d9a2af99b3b" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.744839 4862 scope.go:117] "RemoveContainer" containerID="72dd7e05b0e98629a3547aa5ced1d209fea4cbcd7cb7dc51836836e79ae4b826" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.769921 4862 scope.go:117] "RemoveContainer" containerID="cb861c743ce00570c223afefb40f3cc5ed244b9682a5c2e0bbf669dba92ca34f" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.786307 4862 scope.go:117] "RemoveContainer" containerID="16d4372c7aab7bdaa122d6237f66d4b150a71493de60bf9bed530c808ad0b1a1" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.803617 4862 scope.go:117] "RemoveContainer" containerID="72d37c2517d0efed5caf1404b498bb68525c5f4058d1148ee00d17548c09a99f" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.855640 4862 scope.go:117] "RemoveContainer" containerID="2732ee7a075e6578c8913e90a93812b8f69feb263ac51f3983a02e4f5cb85ed0" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.871563 4862 scope.go:117] "RemoveContainer" containerID="ca4211ba57f4a2e251d0702c6e45025532c02da611b321564e0087fbd70a912b" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.886085 4862 scope.go:117] "RemoveContainer" containerID="36eb5af2c95a5dd49963c18b778f72ccbf03d09076fbe15508a1df0b8532d372" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.907456 4862 scope.go:117] "RemoveContainer" containerID="e2ae775b529fe1449b01a84b2dfacab55da288eac72f7eaca548c563190f06b2" Nov 28 11:51:18 crc kubenswrapper[4862]: I1128 11:51:18.929276 4862 scope.go:117] "RemoveContainer" containerID="32319cdb56353b11cfea025f3670b75833c0097861bc5bbd3e48e49d8ea01757" Nov 28 11:51:30 crc kubenswrapper[4862]: I1128 11:51:30.838935 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:51:30 crc kubenswrapper[4862]: E1128 11:51:30.840167 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:51:45 crc kubenswrapper[4862]: I1128 11:51:45.838893 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:51:45 crc kubenswrapper[4862]: E1128 11:51:45.840124 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:51:56 crc kubenswrapper[4862]: I1128 11:51:56.838789 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:51:56 crc kubenswrapper[4862]: E1128 11:51:56.839575 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:52:08 crc kubenswrapper[4862]: I1128 11:52:08.839816 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:52:08 crc kubenswrapper[4862]: E1128 11:52:08.841236 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:52:19 crc kubenswrapper[4862]: I1128 11:52:19.111315 4862 scope.go:117] "RemoveContainer" containerID="a489662ec3258cf18713de72b1f2ca3ecbbd5ef2ee41d3665720ed4da7de42aa" Nov 28 11:52:19 crc kubenswrapper[4862]: I1128 11:52:19.151444 4862 scope.go:117] "RemoveContainer" containerID="f3dedbee71dddb90736b530eb1e7f1b62741d89591234e177e6f9c26bdde78a3" Nov 28 11:52:19 crc kubenswrapper[4862]: I1128 11:52:19.192445 4862 scope.go:117] "RemoveContainer" containerID="13ea66c81ff754ef5ada0de2b9353279bd3f56a0dc6a739590c85183e537da8e" Nov 28 11:52:19 crc kubenswrapper[4862]: I1128 11:52:19.223036 4862 scope.go:117] "RemoveContainer" containerID="369b84866b935e63fd85b445ef25d538f165e31d60f5a76ac7ab83b20dcfad0e" Nov 28 11:52:19 crc kubenswrapper[4862]: I1128 11:52:19.281802 4862 scope.go:117] "RemoveContainer" containerID="d351e93457fe5a70f380ffce1f54e4bea632ff6f5419983422bc2fc2947f982d" Nov 28 11:52:19 crc kubenswrapper[4862]: I1128 11:52:19.312765 4862 scope.go:117] "RemoveContainer" containerID="61cfcc4fce39fe3c82b8a0422d64efcda70ec22171fe40b6fdd21c27dbcf7bd7" Nov 28 11:52:22 crc kubenswrapper[4862]: I1128 11:52:22.839064 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:52:22 crc kubenswrapper[4862]: E1128 11:52:22.840152 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:52:35 crc kubenswrapper[4862]: I1128 11:52:35.839041 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:52:35 crc kubenswrapper[4862]: E1128 11:52:35.840346 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:52:48 crc kubenswrapper[4862]: I1128 11:52:48.839483 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:52:48 crc kubenswrapper[4862]: E1128 11:52:48.840488 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.937165 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ftbd9/must-gather-cm6q8"] Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938235 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938254 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938273 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerName="neutron-api" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938281 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerName="neutron-api" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938290 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-auditor" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938298 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-auditor" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938312 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-updater" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938319 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-updater" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938328 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="rsync" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938335 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="rsync" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938342 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerName="neutron-httpd" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938349 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerName="neutron-httpd" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938361 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server-init" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938370 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server-init" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938382 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-auditor" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938390 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-auditor" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938398 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938405 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-server" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938414 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-replicator" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938420 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-replicator" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938428 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerName="registry-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938437 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerName="registry-server" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938444 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-reaper" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938451 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-reaper" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938466 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="swift-recon-cron" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938472 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="swift-recon-cron" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938483 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-replicator" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938490 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-replicator" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938500 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="880d930f-ea3a-4a7b-a437-5b164fbea09e" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938508 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="880d930f-ea3a-4a7b-a437-5b164fbea09e" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938524 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed3d880d-6762-4d57-881d-bc585318d500" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938533 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed3d880d-6762-4d57-881d-bc585318d500" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938543 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938549 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938566 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-replicator" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938573 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-replicator" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938583 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938590 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938600 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerName="extract-content" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938606 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerName="extract-content" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938619 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerName="extract-utilities" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938625 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerName="extract-utilities" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938634 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938641 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-server" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938653 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-expirer" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938659 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-expirer" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938672 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71c0d0c9-ed19-41a0-a3d1-92088179bf04" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938679 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="71c0d0c9-ed19-41a0-a3d1-92088179bf04" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938689 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-updater" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938706 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-updater" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938717 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="409c7014-982c-4c60-b374-0ff46c398db4" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938724 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="409c7014-982c-4c60-b374-0ff46c398db4" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938738 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938745 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-server" Nov 28 11:53:02 crc kubenswrapper[4862]: E1128 11:53:02.938755 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-auditor" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938761 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-auditor" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938971 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-replicator" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.938991 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="880d930f-ea3a-4a7b-a437-5b164fbea09e" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939008 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="409c7014-982c-4c60-b374-0ff46c398db4" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939021 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939033 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-auditor" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939045 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-auditor" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939053 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-replicator" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939064 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovs-vswitchd" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939076 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba7cc6b9-1443-45de-84f4-4ec3b92e699c" containerName="ovsdb-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939109 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed3d880d-6762-4d57-881d-bc585318d500" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939123 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939136 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939146 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="8631bc3d-1854-4bfa-b14a-e68dca7567e5" containerName="registry-server" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939154 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-reaper" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939165 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-updater" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939174 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="account-replicator" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939187 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-updater" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939194 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfce9c0c-efa0-4dc0-a7e9-ac7c3c989f9f" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939207 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="container-auditor" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939218 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="object-expirer" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939229 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerName="neutron-api" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939238 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="71c0d0c9-ed19-41a0-a3d1-92088179bf04" containerName="mariadb-account-delete" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939249 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="rsync" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939258 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56372b7-fc39-40cc-8e0a-e95f907db697" containerName="swift-recon-cron" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.939270 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="a42b2a07-23e7-4281-b69b-e47635bcccf6" containerName="neutron-httpd" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.940300 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.950630 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ftbd9/must-gather-cm6q8"] Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.957083 4862 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ftbd9"/"default-dockercfg-5zs4h" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.957358 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ftbd9"/"openshift-service-ca.crt" Nov 28 11:53:02 crc kubenswrapper[4862]: I1128 11:53:02.957519 4862 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ftbd9"/"kube-root-ca.crt" Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.080381 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmdnn\" (UniqueName: \"kubernetes.io/projected/f496ae16-69a3-4b34-a7e8-eabda1679a58-kube-api-access-mmdnn\") pod \"must-gather-cm6q8\" (UID: \"f496ae16-69a3-4b34-a7e8-eabda1679a58\") " pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.080440 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f496ae16-69a3-4b34-a7e8-eabda1679a58-must-gather-output\") pod \"must-gather-cm6q8\" (UID: \"f496ae16-69a3-4b34-a7e8-eabda1679a58\") " pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.181634 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmdnn\" (UniqueName: \"kubernetes.io/projected/f496ae16-69a3-4b34-a7e8-eabda1679a58-kube-api-access-mmdnn\") pod \"must-gather-cm6q8\" (UID: \"f496ae16-69a3-4b34-a7e8-eabda1679a58\") " pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.181856 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f496ae16-69a3-4b34-a7e8-eabda1679a58-must-gather-output\") pod \"must-gather-cm6q8\" (UID: \"f496ae16-69a3-4b34-a7e8-eabda1679a58\") " pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.182283 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f496ae16-69a3-4b34-a7e8-eabda1679a58-must-gather-output\") pod \"must-gather-cm6q8\" (UID: \"f496ae16-69a3-4b34-a7e8-eabda1679a58\") " pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.202959 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmdnn\" (UniqueName: \"kubernetes.io/projected/f496ae16-69a3-4b34-a7e8-eabda1679a58-kube-api-access-mmdnn\") pod \"must-gather-cm6q8\" (UID: \"f496ae16-69a3-4b34-a7e8-eabda1679a58\") " pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.287282 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.731312 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ftbd9/must-gather-cm6q8"] Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.748369 4862 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 11:53:03 crc kubenswrapper[4862]: I1128 11:53:03.838708 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:53:03 crc kubenswrapper[4862]: E1128 11:53:03.838929 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:53:04 crc kubenswrapper[4862]: I1128 11:53:04.489627 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" event={"ID":"f496ae16-69a3-4b34-a7e8-eabda1679a58","Type":"ContainerStarted","Data":"5e94eecaf6376bdac9a511befc30460bb17f0e6c3f24961565057791ba35f62d"} Nov 28 11:53:11 crc kubenswrapper[4862]: I1128 11:53:11.548340 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" event={"ID":"f496ae16-69a3-4b34-a7e8-eabda1679a58","Type":"ContainerStarted","Data":"58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758"} Nov 28 11:53:11 crc kubenswrapper[4862]: I1128 11:53:11.548950 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" event={"ID":"f496ae16-69a3-4b34-a7e8-eabda1679a58","Type":"ContainerStarted","Data":"a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256"} Nov 28 11:53:11 crc kubenswrapper[4862]: I1128 11:53:11.566408 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" podStartSLOduration=2.921809494 podStartE2EDuration="9.566388763s" podCreationTimestamp="2025-11-28 11:53:02 +0000 UTC" firstStartedPulling="2025-11-28 11:53:03.748031966 +0000 UTC m=+1861.280545897" lastFinishedPulling="2025-11-28 11:53:10.392611235 +0000 UTC m=+1867.925125166" observedRunningTime="2025-11-28 11:53:11.559665753 +0000 UTC m=+1869.092179684" watchObservedRunningTime="2025-11-28 11:53:11.566388763 +0000 UTC m=+1869.098902684" Nov 28 11:53:17 crc kubenswrapper[4862]: I1128 11:53:17.839294 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:53:17 crc kubenswrapper[4862]: E1128 11:53:17.839894 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:53:19 crc kubenswrapper[4862]: I1128 11:53:19.429833 4862 scope.go:117] "RemoveContainer" containerID="e3abc0084149b3ca5bb72b8e983fd6de4df9cf6de8997495203d2cf08dd8bfa8" Nov 28 11:53:30 crc kubenswrapper[4862]: I1128 11:53:30.839572 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:53:30 crc kubenswrapper[4862]: E1128 11:53:30.842218 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:53:44 crc kubenswrapper[4862]: I1128 11:53:44.839066 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:53:44 crc kubenswrapper[4862]: E1128 11:53:44.840220 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:53:58 crc kubenswrapper[4862]: I1128 11:53:58.838982 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:53:58 crc kubenswrapper[4862]: E1128 11:53:58.840282 4862 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fm8j2_openshift-machine-config-operator(a6845439-1a4c-4f28-9ba4-29149d545995)\"" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.222689 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6_d23c71bc-2135-4017-a1f7-dabbf039770b/util/0.log" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.420074 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6_d23c71bc-2135-4017-a1f7-dabbf039770b/pull/0.log" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.466691 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6_d23c71bc-2135-4017-a1f7-dabbf039770b/util/0.log" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.470128 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6_d23c71bc-2135-4017-a1f7-dabbf039770b/pull/0.log" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.615974 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6_d23c71bc-2135-4017-a1f7-dabbf039770b/util/0.log" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.633632 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6_d23c71bc-2135-4017-a1f7-dabbf039770b/pull/0.log" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.661556 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aa4msm6_d23c71bc-2135-4017-a1f7-dabbf039770b/extract/0.log" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.794316 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-dj8zz_8bba217a-0d90-4544-b772-62e6e5e58d74/kube-rbac-proxy/0.log" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.838402 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.875289 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-dj8zz_8bba217a-0d90-4544-b772-62e6e5e58d74/manager/0.log" Nov 28 11:54:09 crc kubenswrapper[4862]: I1128 11:54:09.878530 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-7vhdh_f8ce3d3e-999e-48bf-887a-add2b478966d/kube-rbac-proxy/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.027255 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-7vhdh_f8ce3d3e-999e-48bf-887a-add2b478966d/manager/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.044052 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"25337270d83fad17623f37a3faa4c69ac436558797d0d4e3a32bd67f376149f8"} Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.114139 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-5qbwv_b1165cf7-ed1f-48ba-a6dc-1cab53edf166/manager/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.135588 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-5qbwv_b1165cf7-ed1f-48ba-a6dc-1cab53edf166/kube-rbac-proxy/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.275740 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-6h5f6_04fbac3c-ab55-4790-85b2-3269762d6a55/kube-rbac-proxy/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.438085 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-6h5f6_04fbac3c-ab55-4790-85b2-3269762d6a55/manager/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.484688 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-v5t45_50269334-ed51-4bfc-8b87-1243d75774a3/kube-rbac-proxy/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.535955 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-v5t45_50269334-ed51-4bfc-8b87-1243d75774a3/manager/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.694161 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-lnvbp_f33dc056-b642-4cf9-95d4-9c40ac0468ca/kube-rbac-proxy/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.695994 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-lnvbp_f33dc056-b642-4cf9-95d4-9c40ac0468ca/manager/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.781998 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-nxpb6_f4857ca4-a829-4205-b776-fa73faf12eac/kube-rbac-proxy/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.952190 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-nxpb6_f4857ca4-a829-4205-b776-fa73faf12eac/manager/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.955223 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-dmqbn_c50b92c1-e947-4f41-abfe-278e1826a0c9/manager/0.log" Nov 28 11:54:10 crc kubenswrapper[4862]: I1128 11:54:10.990608 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-dmqbn_c50b92c1-e947-4f41-abfe-278e1826a0c9/kube-rbac-proxy/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.129524 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-7tqrc_a836e80a-5070-4ac6-9d06-13f7dcb8f5a7/kube-rbac-proxy/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.208111 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-7tqrc_a836e80a-5070-4ac6-9d06-13f7dcb8f5a7/manager/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.321925 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-tmq8w_9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4/kube-rbac-proxy/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.339888 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-tmq8w_9b13966c-bbb1-4f3d-94a9-6ceb843bcfb4/manager/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.420299 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-vhg8w_4c8abe61-f6dd-4bf2-80e6-c13db3db907f/kube-rbac-proxy/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.522430 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-vhg8w_4c8abe61-f6dd-4bf2-80e6-c13db3db907f/manager/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.581354 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-776s4_4ae22c3f-2ad0-4bde-8045-8e30d5d65d64/kube-rbac-proxy/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.673743 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-776s4_4ae22c3f-2ad0-4bde-8045-8e30d5d65d64/manager/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.725126 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-r6stc_a0567ace-33d0-4e27-a4c6-739b24a847c8/kube-rbac-proxy/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.894050 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-r6stc_a0567ace-33d0-4e27-a4c6-739b24a847c8/manager/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.918813 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-xsqj2_5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56/kube-rbac-proxy/0.log" Nov 28 11:54:11 crc kubenswrapper[4862]: I1128 11:54:11.957055 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-xsqj2_5d3c66e2-5a0f-47fb-a2e2-6acb8aab6f56/manager/0.log" Nov 28 11:54:12 crc kubenswrapper[4862]: I1128 11:54:12.096403 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6c5cc85f67glltc_36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e/kube-rbac-proxy/0.log" Nov 28 11:54:12 crc kubenswrapper[4862]: I1128 11:54:12.143472 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6c5cc85f67glltc_36af3eb7-8fd0-41b5-a7f7-8b42ecdecb8e/manager/0.log" Nov 28 11:54:12 crc kubenswrapper[4862]: I1128 11:54:12.571653 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-g64fp_e37fdd75-4a2c-4758-a2de-9dd365f7ead6/registry-server/0.log" Nov 28 11:54:12 crc kubenswrapper[4862]: I1128 11:54:12.594431 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-848c47db98-xgf7q_dd995aaf-688a-4234-8132-b2ea19bd9fff/operator/0.log" Nov 28 11:54:12 crc kubenswrapper[4862]: I1128 11:54:12.773944 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-54k27_be84305c-9d9e-425d-a8d3-f1ce56254552/kube-rbac-proxy/0.log" Nov 28 11:54:12 crc kubenswrapper[4862]: I1128 11:54:12.849456 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-54k27_be84305c-9d9e-425d-a8d3-f1ce56254552/manager/0.log" Nov 28 11:54:12 crc kubenswrapper[4862]: I1128 11:54:12.964321 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-n2hrf_84002eaf-76bf-4a7e-a7ef-81c87d450270/kube-rbac-proxy/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.026037 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-n2hrf_84002eaf-76bf-4a7e-a7ef-81c87d450270/manager/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.091196 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-xtc9t_50d316aa-a1e4-4ca0-907a-52537f1b1218/operator/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.123951 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6f5f885fb5-zf5gf_b19e018e-f54a-4f60-bef9-345839ae970c/manager/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.237809 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-jz7cq_9e68a5c6-829c-4897-b251-c181097105fc/kube-rbac-proxy/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.273277 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-jz7cq_9e68a5c6-829c-4897-b251-c181097105fc/manager/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.358823 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-trj7j_d8d42319-792f-4aec-abcc-7158a1bf9f9c/kube-rbac-proxy/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.437026 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-trj7j_d8d42319-792f-4aec-abcc-7158a1bf9f9c/manager/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.526240 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-d4bgt_176a3ccc-ec57-4af5-9e09-cb958ed71a2b/kube-rbac-proxy/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.526605 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-d4bgt_176a3ccc-ec57-4af5-9e09-cb958ed71a2b/manager/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.644248 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-p5n9j_70903410-bdee-42fc-a1bf-2213f259dd6f/kube-rbac-proxy/0.log" Nov 28 11:54:13 crc kubenswrapper[4862]: I1128 11:54:13.679503 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-p5n9j_70903410-bdee-42fc-a1bf-2213f259dd6f/manager/0.log" Nov 28 11:54:19 crc kubenswrapper[4862]: I1128 11:54:19.503879 4862 scope.go:117] "RemoveContainer" containerID="2b26e63131e439c21df4fae7d3c514f79108333715e2829d1a0fb8c5a22b7e06" Nov 28 11:54:19 crc kubenswrapper[4862]: I1128 11:54:19.537399 4862 scope.go:117] "RemoveContainer" containerID="37b2c7afc1b4ec62391df0e41265269aef551c36d04b2081170ece771b12ffde" Nov 28 11:54:31 crc kubenswrapper[4862]: I1128 11:54:31.622960 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9xwmh_47f57404-6c91-49f7-bfc5-8a4c2ddaca35/control-plane-machine-set-operator/0.log" Nov 28 11:54:31 crc kubenswrapper[4862]: I1128 11:54:31.733150 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qlr6s_3ad0f875-1ae0-4623-807e-8e88031bbd68/kube-rbac-proxy/0.log" Nov 28 11:54:31 crc kubenswrapper[4862]: I1128 11:54:31.841056 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qlr6s_3ad0f875-1ae0-4623-807e-8e88031bbd68/machine-api-operator/0.log" Nov 28 11:54:45 crc kubenswrapper[4862]: I1128 11:54:45.401610 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-cml9m_43876f80-dd71-47f5-8b77-f22fb71e23f4/cert-manager-controller/0.log" Nov 28 11:54:45 crc kubenswrapper[4862]: I1128 11:54:45.555727 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-xrhk5_5d559bcd-98f2-43ea-9f49-e2a2a8d39c32/cert-manager-cainjector/0.log" Nov 28 11:54:45 crc kubenswrapper[4862]: I1128 11:54:45.580843 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-d9zx4_a59357d2-2d8f-4701-a716-2253d0f35786/cert-manager-webhook/0.log" Nov 28 11:54:59 crc kubenswrapper[4862]: I1128 11:54:59.828265 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-k65jk_49d902d7-ba96-4136-ab9b-e544620bb88a/nmstate-console-plugin/0.log" Nov 28 11:55:00 crc kubenswrapper[4862]: I1128 11:55:00.001592 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gn2tn_3ff44fb9-0a14-4ec9-a5bf-0410a47d9a11/nmstate-handler/0.log" Nov 28 11:55:00 crc kubenswrapper[4862]: I1128 11:55:00.001813 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-pngdc_bf1dc471-6ca1-44fb-9c66-d905ad48ac57/kube-rbac-proxy/0.log" Nov 28 11:55:00 crc kubenswrapper[4862]: I1128 11:55:00.006047 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-pngdc_bf1dc471-6ca1-44fb-9c66-d905ad48ac57/nmstate-metrics/0.log" Nov 28 11:55:00 crc kubenswrapper[4862]: I1128 11:55:00.173114 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-jlnt9_d26ccddf-e0a8-474b-87e3-2c9437e8f2eb/nmstate-operator/0.log" Nov 28 11:55:00 crc kubenswrapper[4862]: I1128 11:55:00.225071 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-vtpdb_ac91e27c-fbee-4e6a-9ae1-001d68ada16d/nmstate-webhook/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.265504 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-c7z9m_df58eb63-bc43-4d48-a2ce-d69a3bf053ef/kube-rbac-proxy/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.510783 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-frr-files/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.566474 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-c7z9m_df58eb63-bc43-4d48-a2ce-d69a3bf053ef/controller/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.692933 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-metrics/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.712957 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-reloader/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.724532 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-frr-files/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.735868 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-reloader/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.927392 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-reloader/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.935853 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-frr-files/0.log" Nov 28 11:55:15 crc kubenswrapper[4862]: I1128 11:55:15.958771 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-metrics/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.026143 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-metrics/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.128026 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-frr-files/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.155063 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-reloader/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.167700 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/cp-metrics/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.211187 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/controller/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.341670 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/frr-metrics/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.372461 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/kube-rbac-proxy/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.434487 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/kube-rbac-proxy-frr/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.589701 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/reloader/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.673751 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-7w8vv_01f38525-c12e-4706-b45c-5cba52dd6d02/frr-k8s-webhook-server/0.log" Nov 28 11:55:16 crc kubenswrapper[4862]: I1128 11:55:16.797391 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-78fccdcc4f-7d2dp_aeebe377-e303-4234-9fcc-c43d7d22f738/manager/0.log" Nov 28 11:55:17 crc kubenswrapper[4862]: I1128 11:55:17.052045 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-f7f698684-bhd4s_9fb4f150-6dfc-4aed-a8c6-3c0ba89d39d7/webhook-server/0.log" Nov 28 11:55:17 crc kubenswrapper[4862]: I1128 11:55:17.165765 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qchmm_40c6570e-6b5e-481f-8b37-45e27aee43be/kube-rbac-proxy/0.log" Nov 28 11:55:17 crc kubenswrapper[4862]: I1128 11:55:17.196613 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tz4mw_ce782165-5659-4d92-b449-78a9cd04a9b2/frr/0.log" Nov 28 11:55:17 crc kubenswrapper[4862]: I1128 11:55:17.511595 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qchmm_40c6570e-6b5e-481f-8b37-45e27aee43be/speaker/0.log" Nov 28 11:55:31 crc kubenswrapper[4862]: I1128 11:55:31.287352 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq_95091706-4117-46dc-9405-f633f5467383/util/0.log" Nov 28 11:55:31 crc kubenswrapper[4862]: I1128 11:55:31.491291 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq_95091706-4117-46dc-9405-f633f5467383/util/0.log" Nov 28 11:55:31 crc kubenswrapper[4862]: I1128 11:55:31.504222 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq_95091706-4117-46dc-9405-f633f5467383/pull/0.log" Nov 28 11:55:31 crc kubenswrapper[4862]: I1128 11:55:31.516244 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq_95091706-4117-46dc-9405-f633f5467383/pull/0.log" Nov 28 11:55:31 crc kubenswrapper[4862]: I1128 11:55:31.688426 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq_95091706-4117-46dc-9405-f633f5467383/pull/0.log" Nov 28 11:55:31 crc kubenswrapper[4862]: I1128 11:55:31.688914 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq_95091706-4117-46dc-9405-f633f5467383/util/0.log" Nov 28 11:55:31 crc kubenswrapper[4862]: I1128 11:55:31.715419 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlbzq_95091706-4117-46dc-9405-f633f5467383/extract/0.log" Nov 28 11:55:31 crc kubenswrapper[4862]: I1128 11:55:31.878070 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4_e7347f1d-97ee-4391-9d84-08348032071c/util/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.050507 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4_e7347f1d-97ee-4391-9d84-08348032071c/util/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.052505 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4_e7347f1d-97ee-4391-9d84-08348032071c/pull/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.091414 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4_e7347f1d-97ee-4391-9d84-08348032071c/pull/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.242435 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4_e7347f1d-97ee-4391-9d84-08348032071c/util/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.271905 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4_e7347f1d-97ee-4391-9d84-08348032071c/pull/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.277933 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxz4t4_e7347f1d-97ee-4391-9d84-08348032071c/extract/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.424965 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p_9726dcde-7041-402a-9ab8-9a7b04e3bff6/util/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.593993 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p_9726dcde-7041-402a-9ab8-9a7b04e3bff6/pull/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.604064 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p_9726dcde-7041-402a-9ab8-9a7b04e3bff6/util/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.629371 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p_9726dcde-7041-402a-9ab8-9a7b04e3bff6/pull/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.746652 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p_9726dcde-7041-402a-9ab8-9a7b04e3bff6/util/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.777362 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p_9726dcde-7041-402a-9ab8-9a7b04e3bff6/pull/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.780922 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836v67p_9726dcde-7041-402a-9ab8-9a7b04e3bff6/extract/0.log" Nov 28 11:55:32 crc kubenswrapper[4862]: I1128 11:55:32.955997 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sz82_93cd8790-a274-4fdb-8c48-028231776561/extract-utilities/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.142613 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sz82_93cd8790-a274-4fdb-8c48-028231776561/extract-content/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.145717 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sz82_93cd8790-a274-4fdb-8c48-028231776561/extract-utilities/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.206002 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sz82_93cd8790-a274-4fdb-8c48-028231776561/extract-content/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.327709 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sz82_93cd8790-a274-4fdb-8c48-028231776561/extract-utilities/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.384493 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sz82_93cd8790-a274-4fdb-8c48-028231776561/extract-content/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.538405 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q66nw_6a08c877-6f2a-4c68-8764-f9cb39e16387/extract-utilities/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.702873 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sz82_93cd8790-a274-4fdb-8c48-028231776561/registry-server/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.726081 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q66nw_6a08c877-6f2a-4c68-8764-f9cb39e16387/extract-content/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.761759 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q66nw_6a08c877-6f2a-4c68-8764-f9cb39e16387/extract-utilities/0.log" Nov 28 11:55:33 crc kubenswrapper[4862]: I1128 11:55:33.773842 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q66nw_6a08c877-6f2a-4c68-8764-f9cb39e16387/extract-content/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.023257 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q66nw_6a08c877-6f2a-4c68-8764-f9cb39e16387/extract-content/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.043709 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q66nw_6a08c877-6f2a-4c68-8764-f9cb39e16387/extract-utilities/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.255845 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q7zkb_db7b8556-bd2d-4ed1-b3c2-984e207ae30b/marketplace-operator/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.327982 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f8552_67725ff2-31c1-49ea-895e-36734cdab9ef/extract-utilities/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.441858 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q66nw_6a08c877-6f2a-4c68-8764-f9cb39e16387/registry-server/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.543115 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f8552_67725ff2-31c1-49ea-895e-36734cdab9ef/extract-utilities/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.548557 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f8552_67725ff2-31c1-49ea-895e-36734cdab9ef/extract-content/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.548657 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f8552_67725ff2-31c1-49ea-895e-36734cdab9ef/extract-content/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.762372 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f8552_67725ff2-31c1-49ea-895e-36734cdab9ef/extract-content/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.766772 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f8552_67725ff2-31c1-49ea-895e-36734cdab9ef/extract-utilities/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.786516 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f8552_67725ff2-31c1-49ea-895e-36734cdab9ef/registry-server/0.log" Nov 28 11:55:34 crc kubenswrapper[4862]: I1128 11:55:34.863388 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dwqgj_de97f27c-53ed-42fa-ba3f-4fd817a2e32f/extract-utilities/0.log" Nov 28 11:55:35 crc kubenswrapper[4862]: I1128 11:55:35.061464 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dwqgj_de97f27c-53ed-42fa-ba3f-4fd817a2e32f/extract-utilities/0.log" Nov 28 11:55:35 crc kubenswrapper[4862]: I1128 11:55:35.075535 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dwqgj_de97f27c-53ed-42fa-ba3f-4fd817a2e32f/extract-content/0.log" Nov 28 11:55:35 crc kubenswrapper[4862]: I1128 11:55:35.092028 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dwqgj_de97f27c-53ed-42fa-ba3f-4fd817a2e32f/extract-content/0.log" Nov 28 11:55:35 crc kubenswrapper[4862]: I1128 11:55:35.243986 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dwqgj_de97f27c-53ed-42fa-ba3f-4fd817a2e32f/extract-utilities/0.log" Nov 28 11:55:35 crc kubenswrapper[4862]: I1128 11:55:35.254141 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dwqgj_de97f27c-53ed-42fa-ba3f-4fd817a2e32f/extract-content/0.log" Nov 28 11:55:35 crc kubenswrapper[4862]: I1128 11:55:35.484039 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dwqgj_de97f27c-53ed-42fa-ba3f-4fd817a2e32f/registry-server/0.log" Nov 28 11:56:38 crc kubenswrapper[4862]: I1128 11:56:38.292082 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:56:38 crc kubenswrapper[4862]: I1128 11:56:38.292773 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:56:42 crc kubenswrapper[4862]: I1128 11:56:42.731304 4862 generic.go:334] "Generic (PLEG): container finished" podID="f496ae16-69a3-4b34-a7e8-eabda1679a58" containerID="a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256" exitCode=0 Nov 28 11:56:42 crc kubenswrapper[4862]: I1128 11:56:42.731507 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" event={"ID":"f496ae16-69a3-4b34-a7e8-eabda1679a58","Type":"ContainerDied","Data":"a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256"} Nov 28 11:56:42 crc kubenswrapper[4862]: I1128 11:56:42.732712 4862 scope.go:117] "RemoveContainer" containerID="a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256" Nov 28 11:56:43 crc kubenswrapper[4862]: I1128 11:56:43.694308 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ftbd9_must-gather-cm6q8_f496ae16-69a3-4b34-a7e8-eabda1679a58/gather/0.log" Nov 28 11:56:49 crc kubenswrapper[4862]: I1128 11:56:49.943673 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l259b"] Nov 28 11:56:49 crc kubenswrapper[4862]: I1128 11:56:49.945659 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:49 crc kubenswrapper[4862]: I1128 11:56:49.956560 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l259b"] Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.080658 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-catalog-content\") pod \"redhat-operators-l259b\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.080957 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zk27\" (UniqueName: \"kubernetes.io/projected/e581a60b-cf3f-4c28-b458-8230bc8ff631-kube-api-access-9zk27\") pod \"redhat-operators-l259b\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.081047 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-utilities\") pod \"redhat-operators-l259b\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.182926 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zk27\" (UniqueName: \"kubernetes.io/projected/e581a60b-cf3f-4c28-b458-8230bc8ff631-kube-api-access-9zk27\") pod \"redhat-operators-l259b\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.182980 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-utilities\") pod \"redhat-operators-l259b\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.183054 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-catalog-content\") pod \"redhat-operators-l259b\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.183611 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-utilities\") pod \"redhat-operators-l259b\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.183665 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-catalog-content\") pod \"redhat-operators-l259b\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.207315 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zk27\" (UniqueName: \"kubernetes.io/projected/e581a60b-cf3f-4c28-b458-8230bc8ff631-kube-api-access-9zk27\") pod \"redhat-operators-l259b\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.262749 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.724745 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l259b"] Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.913222 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l259b" event={"ID":"e581a60b-cf3f-4c28-b458-8230bc8ff631","Type":"ContainerStarted","Data":"983c1f61108cae1fe006c41c10dfa53d16bb916623e77effa64e9a4759066dac"} Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.940544 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ftbd9/must-gather-cm6q8"] Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.942347 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" podUID="f496ae16-69a3-4b34-a7e8-eabda1679a58" containerName="copy" containerID="cri-o://58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758" gracePeriod=2 Nov 28 11:56:50 crc kubenswrapper[4862]: I1128 11:56:50.955609 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ftbd9/must-gather-cm6q8"] Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.332285 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ftbd9_must-gather-cm6q8_f496ae16-69a3-4b34-a7e8-eabda1679a58/copy/0.log" Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.332997 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.405013 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f496ae16-69a3-4b34-a7e8-eabda1679a58-must-gather-output\") pod \"f496ae16-69a3-4b34-a7e8-eabda1679a58\" (UID: \"f496ae16-69a3-4b34-a7e8-eabda1679a58\") " Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.405071 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmdnn\" (UniqueName: \"kubernetes.io/projected/f496ae16-69a3-4b34-a7e8-eabda1679a58-kube-api-access-mmdnn\") pod \"f496ae16-69a3-4b34-a7e8-eabda1679a58\" (UID: \"f496ae16-69a3-4b34-a7e8-eabda1679a58\") " Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.458150 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f496ae16-69a3-4b34-a7e8-eabda1679a58-kube-api-access-mmdnn" (OuterVolumeSpecName: "kube-api-access-mmdnn") pod "f496ae16-69a3-4b34-a7e8-eabda1679a58" (UID: "f496ae16-69a3-4b34-a7e8-eabda1679a58"). InnerVolumeSpecName "kube-api-access-mmdnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.509631 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmdnn\" (UniqueName: \"kubernetes.io/projected/f496ae16-69a3-4b34-a7e8-eabda1679a58-kube-api-access-mmdnn\") on node \"crc\" DevicePath \"\"" Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.544136 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f496ae16-69a3-4b34-a7e8-eabda1679a58-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "f496ae16-69a3-4b34-a7e8-eabda1679a58" (UID: "f496ae16-69a3-4b34-a7e8-eabda1679a58"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.611240 4862 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f496ae16-69a3-4b34-a7e8-eabda1679a58-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.923398 4862 generic.go:334] "Generic (PLEG): container finished" podID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerID="571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80" exitCode=0 Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.923488 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l259b" event={"ID":"e581a60b-cf3f-4c28-b458-8230bc8ff631","Type":"ContainerDied","Data":"571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80"} Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.925468 4862 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ftbd9_must-gather-cm6q8_f496ae16-69a3-4b34-a7e8-eabda1679a58/copy/0.log" Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.926045 4862 generic.go:334] "Generic (PLEG): container finished" podID="f496ae16-69a3-4b34-a7e8-eabda1679a58" containerID="58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758" exitCode=143 Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.926139 4862 scope.go:117] "RemoveContainer" containerID="58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758" Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.926146 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ftbd9/must-gather-cm6q8" Nov 28 11:56:51 crc kubenswrapper[4862]: I1128 11:56:51.956267 4862 scope.go:117] "RemoveContainer" containerID="a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256" Nov 28 11:56:52 crc kubenswrapper[4862]: I1128 11:56:52.027648 4862 scope.go:117] "RemoveContainer" containerID="58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758" Nov 28 11:56:52 crc kubenswrapper[4862]: E1128 11:56:52.029087 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758\": container with ID starting with 58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758 not found: ID does not exist" containerID="58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758" Nov 28 11:56:52 crc kubenswrapper[4862]: I1128 11:56:52.029235 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758"} err="failed to get container status \"58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758\": rpc error: code = NotFound desc = could not find container \"58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758\": container with ID starting with 58b261bd920b3210acf33a9aa667937695359cad69d1dc49553f6d8701f9f758 not found: ID does not exist" Nov 28 11:56:52 crc kubenswrapper[4862]: I1128 11:56:52.029339 4862 scope.go:117] "RemoveContainer" containerID="a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256" Nov 28 11:56:52 crc kubenswrapper[4862]: E1128 11:56:52.029889 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256\": container with ID starting with a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256 not found: ID does not exist" containerID="a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256" Nov 28 11:56:52 crc kubenswrapper[4862]: I1128 11:56:52.029953 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256"} err="failed to get container status \"a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256\": rpc error: code = NotFound desc = could not find container \"a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256\": container with ID starting with a1f601e45705496dd996e4b1ffa1df010753399e6b031ccc3ed9dec8c06b9256 not found: ID does not exist" Nov 28 11:56:52 crc kubenswrapper[4862]: I1128 11:56:52.849914 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f496ae16-69a3-4b34-a7e8-eabda1679a58" path="/var/lib/kubelet/pods/f496ae16-69a3-4b34-a7e8-eabda1679a58/volumes" Nov 28 11:56:53 crc kubenswrapper[4862]: I1128 11:56:53.960728 4862 generic.go:334] "Generic (PLEG): container finished" podID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerID="b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602" exitCode=0 Nov 28 11:56:53 crc kubenswrapper[4862]: I1128 11:56:53.961195 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l259b" event={"ID":"e581a60b-cf3f-4c28-b458-8230bc8ff631","Type":"ContainerDied","Data":"b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602"} Nov 28 11:56:54 crc kubenswrapper[4862]: I1128 11:56:54.969867 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l259b" event={"ID":"e581a60b-cf3f-4c28-b458-8230bc8ff631","Type":"ContainerStarted","Data":"387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e"} Nov 28 11:56:54 crc kubenswrapper[4862]: I1128 11:56:54.989818 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l259b" podStartSLOduration=3.366729174 podStartE2EDuration="5.989804747s" podCreationTimestamp="2025-11-28 11:56:49 +0000 UTC" firstStartedPulling="2025-11-28 11:56:51.925734862 +0000 UTC m=+2089.458248783" lastFinishedPulling="2025-11-28 11:56:54.548810435 +0000 UTC m=+2092.081324356" observedRunningTime="2025-11-28 11:56:54.988326681 +0000 UTC m=+2092.520840602" watchObservedRunningTime="2025-11-28 11:56:54.989804747 +0000 UTC m=+2092.522318668" Nov 28 11:57:00 crc kubenswrapper[4862]: I1128 11:57:00.263040 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:57:00 crc kubenswrapper[4862]: I1128 11:57:00.263653 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:57:01 crc kubenswrapper[4862]: I1128 11:57:01.305186 4862 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l259b" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerName="registry-server" probeResult="failure" output=< Nov 28 11:57:01 crc kubenswrapper[4862]: timeout: failed to connect service ":50051" within 1s Nov 28 11:57:01 crc kubenswrapper[4862]: > Nov 28 11:57:08 crc kubenswrapper[4862]: I1128 11:57:08.291453 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:57:08 crc kubenswrapper[4862]: I1128 11:57:08.292176 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:57:10 crc kubenswrapper[4862]: I1128 11:57:10.329310 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:57:10 crc kubenswrapper[4862]: I1128 11:57:10.395889 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:57:10 crc kubenswrapper[4862]: I1128 11:57:10.575836 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l259b"] Nov 28 11:57:12 crc kubenswrapper[4862]: I1128 11:57:12.131072 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l259b" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerName="registry-server" containerID="cri-o://387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e" gracePeriod=2 Nov 28 11:57:12 crc kubenswrapper[4862]: I1128 11:57:12.703111 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:57:12 crc kubenswrapper[4862]: I1128 11:57:12.824591 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zk27\" (UniqueName: \"kubernetes.io/projected/e581a60b-cf3f-4c28-b458-8230bc8ff631-kube-api-access-9zk27\") pod \"e581a60b-cf3f-4c28-b458-8230bc8ff631\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " Nov 28 11:57:12 crc kubenswrapper[4862]: I1128 11:57:12.824774 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-utilities\") pod \"e581a60b-cf3f-4c28-b458-8230bc8ff631\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " Nov 28 11:57:12 crc kubenswrapper[4862]: I1128 11:57:12.824957 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-catalog-content\") pod \"e581a60b-cf3f-4c28-b458-8230bc8ff631\" (UID: \"e581a60b-cf3f-4c28-b458-8230bc8ff631\") " Nov 28 11:57:12 crc kubenswrapper[4862]: I1128 11:57:12.826180 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-utilities" (OuterVolumeSpecName: "utilities") pod "e581a60b-cf3f-4c28-b458-8230bc8ff631" (UID: "e581a60b-cf3f-4c28-b458-8230bc8ff631"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:57:12 crc kubenswrapper[4862]: I1128 11:57:12.830398 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e581a60b-cf3f-4c28-b458-8230bc8ff631-kube-api-access-9zk27" (OuterVolumeSpecName: "kube-api-access-9zk27") pod "e581a60b-cf3f-4c28-b458-8230bc8ff631" (UID: "e581a60b-cf3f-4c28-b458-8230bc8ff631"). InnerVolumeSpecName "kube-api-access-9zk27". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:57:12 crc kubenswrapper[4862]: I1128 11:57:12.926484 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:57:12 crc kubenswrapper[4862]: I1128 11:57:12.926522 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zk27\" (UniqueName: \"kubernetes.io/projected/e581a60b-cf3f-4c28-b458-8230bc8ff631-kube-api-access-9zk27\") on node \"crc\" DevicePath \"\"" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.003320 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e581a60b-cf3f-4c28-b458-8230bc8ff631" (UID: "e581a60b-cf3f-4c28-b458-8230bc8ff631"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.028147 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581a60b-cf3f-4c28-b458-8230bc8ff631-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.141394 4862 generic.go:334] "Generic (PLEG): container finished" podID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerID="387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e" exitCode=0 Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.141443 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l259b" event={"ID":"e581a60b-cf3f-4c28-b458-8230bc8ff631","Type":"ContainerDied","Data":"387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e"} Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.141476 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l259b" event={"ID":"e581a60b-cf3f-4c28-b458-8230bc8ff631","Type":"ContainerDied","Data":"983c1f61108cae1fe006c41c10dfa53d16bb916623e77effa64e9a4759066dac"} Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.141500 4862 scope.go:117] "RemoveContainer" containerID="387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.142693 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l259b" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.193447 4862 scope.go:117] "RemoveContainer" containerID="b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.197328 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l259b"] Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.205281 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l259b"] Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.232163 4862 scope.go:117] "RemoveContainer" containerID="571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.259879 4862 scope.go:117] "RemoveContainer" containerID="387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e" Nov 28 11:57:13 crc kubenswrapper[4862]: E1128 11:57:13.260285 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e\": container with ID starting with 387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e not found: ID does not exist" containerID="387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.260322 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e"} err="failed to get container status \"387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e\": rpc error: code = NotFound desc = could not find container \"387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e\": container with ID starting with 387146de77afb463f3edc0761790cc84ea715731ca8eda6f6a81f0c837ea0e6e not found: ID does not exist" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.260349 4862 scope.go:117] "RemoveContainer" containerID="b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602" Nov 28 11:57:13 crc kubenswrapper[4862]: E1128 11:57:13.261578 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602\": container with ID starting with b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602 not found: ID does not exist" containerID="b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.261612 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602"} err="failed to get container status \"b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602\": rpc error: code = NotFound desc = could not find container \"b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602\": container with ID starting with b42ae274a24bb502f5928b2e80ee5f64b8e2d9edd032f6a58f7eb4cbc08b2602 not found: ID does not exist" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.261648 4862 scope.go:117] "RemoveContainer" containerID="571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80" Nov 28 11:57:13 crc kubenswrapper[4862]: E1128 11:57:13.261923 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80\": container with ID starting with 571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80 not found: ID does not exist" containerID="571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80" Nov 28 11:57:13 crc kubenswrapper[4862]: I1128 11:57:13.261957 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80"} err="failed to get container status \"571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80\": rpc error: code = NotFound desc = could not find container \"571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80\": container with ID starting with 571872df2654e32f86c4f8076cd6cda91e7a2fbe3872cc1704870eea0c3e3d80 not found: ID does not exist" Nov 28 11:57:14 crc kubenswrapper[4862]: I1128 11:57:14.852786 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" path="/var/lib/kubelet/pods/e581a60b-cf3f-4c28-b458-8230bc8ff631/volumes" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.324648 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k5mlh"] Nov 28 11:57:19 crc kubenswrapper[4862]: E1128 11:57:19.325473 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f496ae16-69a3-4b34-a7e8-eabda1679a58" containerName="copy" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.325489 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f496ae16-69a3-4b34-a7e8-eabda1679a58" containerName="copy" Nov 28 11:57:19 crc kubenswrapper[4862]: E1128 11:57:19.325522 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerName="extract-utilities" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.325530 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerName="extract-utilities" Nov 28 11:57:19 crc kubenswrapper[4862]: E1128 11:57:19.325548 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f496ae16-69a3-4b34-a7e8-eabda1679a58" containerName="gather" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.325558 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="f496ae16-69a3-4b34-a7e8-eabda1679a58" containerName="gather" Nov 28 11:57:19 crc kubenswrapper[4862]: E1128 11:57:19.325573 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerName="extract-content" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.325581 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerName="extract-content" Nov 28 11:57:19 crc kubenswrapper[4862]: E1128 11:57:19.325597 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerName="registry-server" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.325604 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerName="registry-server" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.325766 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f496ae16-69a3-4b34-a7e8-eabda1679a58" containerName="copy" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.325794 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="f496ae16-69a3-4b34-a7e8-eabda1679a58" containerName="gather" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.325807 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="e581a60b-cf3f-4c28-b458-8230bc8ff631" containerName="registry-server" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.326976 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.341119 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k5mlh"] Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.426497 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-catalog-content\") pod \"certified-operators-k5mlh\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.426571 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-utilities\") pod \"certified-operators-k5mlh\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.426605 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd9jd\" (UniqueName: \"kubernetes.io/projected/815b53cc-ef46-401c-8483-0a673458542c-kube-api-access-xd9jd\") pod \"certified-operators-k5mlh\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.528551 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-catalog-content\") pod \"certified-operators-k5mlh\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.528606 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-utilities\") pod \"certified-operators-k5mlh\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.528626 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd9jd\" (UniqueName: \"kubernetes.io/projected/815b53cc-ef46-401c-8483-0a673458542c-kube-api-access-xd9jd\") pod \"certified-operators-k5mlh\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.529082 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-catalog-content\") pod \"certified-operators-k5mlh\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.529165 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-utilities\") pod \"certified-operators-k5mlh\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.545740 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd9jd\" (UniqueName: \"kubernetes.io/projected/815b53cc-ef46-401c-8483-0a673458542c-kube-api-access-xd9jd\") pod \"certified-operators-k5mlh\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:19 crc kubenswrapper[4862]: I1128 11:57:19.660984 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:20 crc kubenswrapper[4862]: I1128 11:57:20.133189 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k5mlh"] Nov 28 11:57:20 crc kubenswrapper[4862]: I1128 11:57:20.220023 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5mlh" event={"ID":"815b53cc-ef46-401c-8483-0a673458542c","Type":"ContainerStarted","Data":"21547e7ec11bb261a49e28d3ecb54e236ab4dfd0f09f56a9e2722be331876cab"} Nov 28 11:57:21 crc kubenswrapper[4862]: I1128 11:57:21.232646 4862 generic.go:334] "Generic (PLEG): container finished" podID="815b53cc-ef46-401c-8483-0a673458542c" containerID="6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294" exitCode=0 Nov 28 11:57:21 crc kubenswrapper[4862]: I1128 11:57:21.232848 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5mlh" event={"ID":"815b53cc-ef46-401c-8483-0a673458542c","Type":"ContainerDied","Data":"6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294"} Nov 28 11:57:23 crc kubenswrapper[4862]: I1128 11:57:23.252704 4862 generic.go:334] "Generic (PLEG): container finished" podID="815b53cc-ef46-401c-8483-0a673458542c" containerID="354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea" exitCode=0 Nov 28 11:57:23 crc kubenswrapper[4862]: I1128 11:57:23.252787 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5mlh" event={"ID":"815b53cc-ef46-401c-8483-0a673458542c","Type":"ContainerDied","Data":"354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea"} Nov 28 11:57:24 crc kubenswrapper[4862]: I1128 11:57:24.264778 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5mlh" event={"ID":"815b53cc-ef46-401c-8483-0a673458542c","Type":"ContainerStarted","Data":"a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634"} Nov 28 11:57:24 crc kubenswrapper[4862]: I1128 11:57:24.290058 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k5mlh" podStartSLOduration=2.832473084 podStartE2EDuration="5.290037438s" podCreationTimestamp="2025-11-28 11:57:19 +0000 UTC" firstStartedPulling="2025-11-28 11:57:21.234977222 +0000 UTC m=+2118.767491173" lastFinishedPulling="2025-11-28 11:57:23.692541606 +0000 UTC m=+2121.225055527" observedRunningTime="2025-11-28 11:57:24.288454638 +0000 UTC m=+2121.820968569" watchObservedRunningTime="2025-11-28 11:57:24.290037438 +0000 UTC m=+2121.822551359" Nov 28 11:57:29 crc kubenswrapper[4862]: I1128 11:57:29.662273 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:29 crc kubenswrapper[4862]: I1128 11:57:29.663235 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:29 crc kubenswrapper[4862]: I1128 11:57:29.747400 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:30 crc kubenswrapper[4862]: I1128 11:57:30.395270 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:30 crc kubenswrapper[4862]: I1128 11:57:30.461022 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k5mlh"] Nov 28 11:57:32 crc kubenswrapper[4862]: I1128 11:57:32.340264 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k5mlh" podUID="815b53cc-ef46-401c-8483-0a673458542c" containerName="registry-server" containerID="cri-o://a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634" gracePeriod=2 Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.293704 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.339736 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd9jd\" (UniqueName: \"kubernetes.io/projected/815b53cc-ef46-401c-8483-0a673458542c-kube-api-access-xd9jd\") pod \"815b53cc-ef46-401c-8483-0a673458542c\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.339829 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-catalog-content\") pod \"815b53cc-ef46-401c-8483-0a673458542c\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.339971 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-utilities\") pod \"815b53cc-ef46-401c-8483-0a673458542c\" (UID: \"815b53cc-ef46-401c-8483-0a673458542c\") " Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.341512 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-utilities" (OuterVolumeSpecName: "utilities") pod "815b53cc-ef46-401c-8483-0a673458542c" (UID: "815b53cc-ef46-401c-8483-0a673458542c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.348646 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/815b53cc-ef46-401c-8483-0a673458542c-kube-api-access-xd9jd" (OuterVolumeSpecName: "kube-api-access-xd9jd") pod "815b53cc-ef46-401c-8483-0a673458542c" (UID: "815b53cc-ef46-401c-8483-0a673458542c"). InnerVolumeSpecName "kube-api-access-xd9jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.372360 4862 generic.go:334] "Generic (PLEG): container finished" podID="815b53cc-ef46-401c-8483-0a673458542c" containerID="a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634" exitCode=0 Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.372405 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5mlh" event={"ID":"815b53cc-ef46-401c-8483-0a673458542c","Type":"ContainerDied","Data":"a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634"} Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.372433 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5mlh" event={"ID":"815b53cc-ef46-401c-8483-0a673458542c","Type":"ContainerDied","Data":"21547e7ec11bb261a49e28d3ecb54e236ab4dfd0f09f56a9e2722be331876cab"} Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.372448 4862 scope.go:117] "RemoveContainer" containerID="a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.372559 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5mlh" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.398850 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "815b53cc-ef46-401c-8483-0a673458542c" (UID: "815b53cc-ef46-401c-8483-0a673458542c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.403871 4862 scope.go:117] "RemoveContainer" containerID="354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.428109 4862 scope.go:117] "RemoveContainer" containerID="6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.442026 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.442066 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd9jd\" (UniqueName: \"kubernetes.io/projected/815b53cc-ef46-401c-8483-0a673458542c-kube-api-access-xd9jd\") on node \"crc\" DevicePath \"\"" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.442080 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815b53cc-ef46-401c-8483-0a673458542c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.455359 4862 scope.go:117] "RemoveContainer" containerID="a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634" Nov 28 11:57:33 crc kubenswrapper[4862]: E1128 11:57:33.455895 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634\": container with ID starting with a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634 not found: ID does not exist" containerID="a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.455926 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634"} err="failed to get container status \"a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634\": rpc error: code = NotFound desc = could not find container \"a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634\": container with ID starting with a7d7ee305c8741a5fba79a2825a9e88d0f0a7be3714b3aff06400c6134d16634 not found: ID does not exist" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.455951 4862 scope.go:117] "RemoveContainer" containerID="354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea" Nov 28 11:57:33 crc kubenswrapper[4862]: E1128 11:57:33.456260 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea\": container with ID starting with 354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea not found: ID does not exist" containerID="354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.456279 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea"} err="failed to get container status \"354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea\": rpc error: code = NotFound desc = could not find container \"354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea\": container with ID starting with 354fc37fef091707df5c0c4bf30e8a8eb4e2b509e338d5fd8a809b64b731fbea not found: ID does not exist" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.456291 4862 scope.go:117] "RemoveContainer" containerID="6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294" Nov 28 11:57:33 crc kubenswrapper[4862]: E1128 11:57:33.456625 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294\": container with ID starting with 6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294 not found: ID does not exist" containerID="6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.456704 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294"} err="failed to get container status \"6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294\": rpc error: code = NotFound desc = could not find container \"6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294\": container with ID starting with 6d3bf18a55ed495e4fc2113d8ceb03d8641e880dfafa13bf393fa24976925294 not found: ID does not exist" Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.708413 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k5mlh"] Nov 28 11:57:33 crc kubenswrapper[4862]: I1128 11:57:33.719736 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k5mlh"] Nov 28 11:57:34 crc kubenswrapper[4862]: I1128 11:57:34.859665 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="815b53cc-ef46-401c-8483-0a673458542c" path="/var/lib/kubelet/pods/815b53cc-ef46-401c-8483-0a673458542c/volumes" Nov 28 11:57:38 crc kubenswrapper[4862]: I1128 11:57:38.291632 4862 patch_prober.go:28] interesting pod/machine-config-daemon-fm8j2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:57:38 crc kubenswrapper[4862]: I1128 11:57:38.293302 4862 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:57:38 crc kubenswrapper[4862]: I1128 11:57:38.293403 4862 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" Nov 28 11:57:38 crc kubenswrapper[4862]: I1128 11:57:38.294503 4862 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25337270d83fad17623f37a3faa4c69ac436558797d0d4e3a32bd67f376149f8"} pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:57:38 crc kubenswrapper[4862]: I1128 11:57:38.294623 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" podUID="a6845439-1a4c-4f28-9ba4-29149d545995" containerName="machine-config-daemon" containerID="cri-o://25337270d83fad17623f37a3faa4c69ac436558797d0d4e3a32bd67f376149f8" gracePeriod=600 Nov 28 11:57:39 crc kubenswrapper[4862]: I1128 11:57:39.431542 4862 generic.go:334] "Generic (PLEG): container finished" podID="a6845439-1a4c-4f28-9ba4-29149d545995" containerID="25337270d83fad17623f37a3faa4c69ac436558797d0d4e3a32bd67f376149f8" exitCode=0 Nov 28 11:57:39 crc kubenswrapper[4862]: I1128 11:57:39.431665 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerDied","Data":"25337270d83fad17623f37a3faa4c69ac436558797d0d4e3a32bd67f376149f8"} Nov 28 11:57:39 crc kubenswrapper[4862]: I1128 11:57:39.431910 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fm8j2" event={"ID":"a6845439-1a4c-4f28-9ba4-29149d545995","Type":"ContainerStarted","Data":"95ff7a23d0a39f12a4818cca2d6c420911c9c554f9e21535e68efcbf28af6d0f"} Nov 28 11:57:39 crc kubenswrapper[4862]: I1128 11:57:39.431940 4862 scope.go:117] "RemoveContainer" containerID="990ac1807ca0e94909f20d1da3bff0e8e91f78771ce439ece5d58710923b9776" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.364876 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-464kv"] Nov 28 11:57:48 crc kubenswrapper[4862]: E1128 11:57:48.366083 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815b53cc-ef46-401c-8483-0a673458542c" containerName="extract-utilities" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.366135 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="815b53cc-ef46-401c-8483-0a673458542c" containerName="extract-utilities" Nov 28 11:57:48 crc kubenswrapper[4862]: E1128 11:57:48.366163 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815b53cc-ef46-401c-8483-0a673458542c" containerName="extract-content" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.366175 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="815b53cc-ef46-401c-8483-0a673458542c" containerName="extract-content" Nov 28 11:57:48 crc kubenswrapper[4862]: E1128 11:57:48.366204 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815b53cc-ef46-401c-8483-0a673458542c" containerName="registry-server" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.366216 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="815b53cc-ef46-401c-8483-0a673458542c" containerName="registry-server" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.366449 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="815b53cc-ef46-401c-8483-0a673458542c" containerName="registry-server" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.367998 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.387478 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-464kv"] Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.468031 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-catalog-content\") pod \"community-operators-464kv\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.468171 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-utilities\") pod \"community-operators-464kv\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.468225 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glpj8\" (UniqueName: \"kubernetes.io/projected/07acf715-5134-4cac-a8d9-289eeb627ccd-kube-api-access-glpj8\") pod \"community-operators-464kv\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.570052 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-utilities\") pod \"community-operators-464kv\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.570140 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glpj8\" (UniqueName: \"kubernetes.io/projected/07acf715-5134-4cac-a8d9-289eeb627ccd-kube-api-access-glpj8\") pod \"community-operators-464kv\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.570207 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-catalog-content\") pod \"community-operators-464kv\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.570924 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-utilities\") pod \"community-operators-464kv\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.570966 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-catalog-content\") pod \"community-operators-464kv\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.603939 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glpj8\" (UniqueName: \"kubernetes.io/projected/07acf715-5134-4cac-a8d9-289eeb627ccd-kube-api-access-glpj8\") pod \"community-operators-464kv\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:48 crc kubenswrapper[4862]: I1128 11:57:48.710756 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:49 crc kubenswrapper[4862]: I1128 11:57:49.211555 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-464kv"] Nov 28 11:57:49 crc kubenswrapper[4862]: I1128 11:57:49.516411 4862 generic.go:334] "Generic (PLEG): container finished" podID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerID="fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a" exitCode=0 Nov 28 11:57:49 crc kubenswrapper[4862]: I1128 11:57:49.516453 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-464kv" event={"ID":"07acf715-5134-4cac-a8d9-289eeb627ccd","Type":"ContainerDied","Data":"fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a"} Nov 28 11:57:49 crc kubenswrapper[4862]: I1128 11:57:49.516485 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-464kv" event={"ID":"07acf715-5134-4cac-a8d9-289eeb627ccd","Type":"ContainerStarted","Data":"8cb9798ee9d78288e7a9e83452469143997c037786e1d50cac0284c1afef9302"} Nov 28 11:57:51 crc kubenswrapper[4862]: I1128 11:57:51.534832 4862 generic.go:334] "Generic (PLEG): container finished" podID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerID="9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08" exitCode=0 Nov 28 11:57:51 crc kubenswrapper[4862]: I1128 11:57:51.534911 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-464kv" event={"ID":"07acf715-5134-4cac-a8d9-289eeb627ccd","Type":"ContainerDied","Data":"9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08"} Nov 28 11:57:52 crc kubenswrapper[4862]: I1128 11:57:52.545474 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-464kv" event={"ID":"07acf715-5134-4cac-a8d9-289eeb627ccd","Type":"ContainerStarted","Data":"e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3"} Nov 28 11:57:52 crc kubenswrapper[4862]: I1128 11:57:52.561808 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-464kv" podStartSLOduration=2.0440961 podStartE2EDuration="4.561787182s" podCreationTimestamp="2025-11-28 11:57:48 +0000 UTC" firstStartedPulling="2025-11-28 11:57:49.520263294 +0000 UTC m=+2147.052777215" lastFinishedPulling="2025-11-28 11:57:52.037954376 +0000 UTC m=+2149.570468297" observedRunningTime="2025-11-28 11:57:52.55996283 +0000 UTC m=+2150.092476751" watchObservedRunningTime="2025-11-28 11:57:52.561787182 +0000 UTC m=+2150.094301113" Nov 28 11:57:58 crc kubenswrapper[4862]: I1128 11:57:58.711125 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:58 crc kubenswrapper[4862]: I1128 11:57:58.711862 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:58 crc kubenswrapper[4862]: I1128 11:57:58.777015 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:59 crc kubenswrapper[4862]: I1128 11:57:59.674715 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-464kv" Nov 28 11:57:59 crc kubenswrapper[4862]: I1128 11:57:59.727271 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-464kv"] Nov 28 11:58:01 crc kubenswrapper[4862]: I1128 11:58:01.619198 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-464kv" podUID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerName="registry-server" containerID="cri-o://e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3" gracePeriod=2 Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.565565 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-464kv" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.612644 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-utilities\") pod \"07acf715-5134-4cac-a8d9-289eeb627ccd\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.612710 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-catalog-content\") pod \"07acf715-5134-4cac-a8d9-289eeb627ccd\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.612769 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glpj8\" (UniqueName: \"kubernetes.io/projected/07acf715-5134-4cac-a8d9-289eeb627ccd-kube-api-access-glpj8\") pod \"07acf715-5134-4cac-a8d9-289eeb627ccd\" (UID: \"07acf715-5134-4cac-a8d9-289eeb627ccd\") " Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.613604 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-utilities" (OuterVolumeSpecName: "utilities") pod "07acf715-5134-4cac-a8d9-289eeb627ccd" (UID: "07acf715-5134-4cac-a8d9-289eeb627ccd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.617947 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07acf715-5134-4cac-a8d9-289eeb627ccd-kube-api-access-glpj8" (OuterVolumeSpecName: "kube-api-access-glpj8") pod "07acf715-5134-4cac-a8d9-289eeb627ccd" (UID: "07acf715-5134-4cac-a8d9-289eeb627ccd"). InnerVolumeSpecName "kube-api-access-glpj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.628970 4862 generic.go:334] "Generic (PLEG): container finished" podID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerID="e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3" exitCode=0 Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.629008 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-464kv" event={"ID":"07acf715-5134-4cac-a8d9-289eeb627ccd","Type":"ContainerDied","Data":"e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3"} Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.629041 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-464kv" event={"ID":"07acf715-5134-4cac-a8d9-289eeb627ccd","Type":"ContainerDied","Data":"8cb9798ee9d78288e7a9e83452469143997c037786e1d50cac0284c1afef9302"} Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.629058 4862 scope.go:117] "RemoveContainer" containerID="e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.629071 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-464kv" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.669053 4862 scope.go:117] "RemoveContainer" containerID="9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.672803 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07acf715-5134-4cac-a8d9-289eeb627ccd" (UID: "07acf715-5134-4cac-a8d9-289eeb627ccd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.690373 4862 scope.go:117] "RemoveContainer" containerID="fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.714212 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glpj8\" (UniqueName: \"kubernetes.io/projected/07acf715-5134-4cac-a8d9-289eeb627ccd-kube-api-access-glpj8\") on node \"crc\" DevicePath \"\"" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.714247 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.714260 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07acf715-5134-4cac-a8d9-289eeb627ccd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.720285 4862 scope.go:117] "RemoveContainer" containerID="e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3" Nov 28 11:58:02 crc kubenswrapper[4862]: E1128 11:58:02.720933 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3\": container with ID starting with e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3 not found: ID does not exist" containerID="e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.720987 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3"} err="failed to get container status \"e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3\": rpc error: code = NotFound desc = could not find container \"e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3\": container with ID starting with e8aa7f49a7da59fa00e65e6359fcf5b2f1f0cd5a7be18fbe7428c33791b322d3 not found: ID does not exist" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.721014 4862 scope.go:117] "RemoveContainer" containerID="9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08" Nov 28 11:58:02 crc kubenswrapper[4862]: E1128 11:58:02.721355 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08\": container with ID starting with 9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08 not found: ID does not exist" containerID="9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.721398 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08"} err="failed to get container status \"9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08\": rpc error: code = NotFound desc = could not find container \"9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08\": container with ID starting with 9554d5ac8a91e1b67794f3b6e5a52f8e53b06e6d52f1ffe689159a75cc6f4f08 not found: ID does not exist" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.721417 4862 scope.go:117] "RemoveContainer" containerID="fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a" Nov 28 11:58:02 crc kubenswrapper[4862]: E1128 11:58:02.721873 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a\": container with ID starting with fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a not found: ID does not exist" containerID="fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.721912 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a"} err="failed to get container status \"fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a\": rpc error: code = NotFound desc = could not find container \"fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a\": container with ID starting with fff691c072232110b861c57fe5467885f535d8f58f25d95a02e74474d26ff03a not found: ID does not exist" Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.956894 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-464kv"] Nov 28 11:58:02 crc kubenswrapper[4862]: I1128 11:58:02.962637 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-464kv"] Nov 28 11:58:04 crc kubenswrapper[4862]: I1128 11:58:04.848287 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07acf715-5134-4cac-a8d9-289eeb627ccd" path="/var/lib/kubelet/pods/07acf715-5134-4cac-a8d9-289eeb627ccd/volumes" Nov 28 11:58:07 crc kubenswrapper[4862]: E1128 11:58:07.902486 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice/crio-8cb9798ee9d78288e7a9e83452469143997c037786e1d50cac0284c1afef9302\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice\": RecentStats: unable to find data in memory cache]" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.310991 4862 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qxsj5"] Nov 28 11:58:14 crc kubenswrapper[4862]: E1128 11:58:14.312195 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerName="extract-content" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.312387 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerName="extract-content" Nov 28 11:58:14 crc kubenswrapper[4862]: E1128 11:58:14.312461 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerName="registry-server" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.312480 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerName="registry-server" Nov 28 11:58:14 crc kubenswrapper[4862]: E1128 11:58:14.312541 4862 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerName="extract-utilities" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.312556 4862 state_mem.go:107] "Deleted CPUSet assignment" podUID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerName="extract-utilities" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.312861 4862 memory_manager.go:354] "RemoveStaleState removing state" podUID="07acf715-5134-4cac-a8d9-289eeb627ccd" containerName="registry-server" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.317850 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.341808 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxsj5"] Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.495868 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-utilities\") pod \"redhat-marketplace-qxsj5\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.495973 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-catalog-content\") pod \"redhat-marketplace-qxsj5\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.496004 4862 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxgjg\" (UniqueName: \"kubernetes.io/projected/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-kube-api-access-lxgjg\") pod \"redhat-marketplace-qxsj5\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.597861 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-catalog-content\") pod \"redhat-marketplace-qxsj5\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.597910 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxgjg\" (UniqueName: \"kubernetes.io/projected/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-kube-api-access-lxgjg\") pod \"redhat-marketplace-qxsj5\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.597990 4862 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-utilities\") pod \"redhat-marketplace-qxsj5\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.598551 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-utilities\") pod \"redhat-marketplace-qxsj5\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.598550 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-catalog-content\") pod \"redhat-marketplace-qxsj5\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.615998 4862 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxgjg\" (UniqueName: \"kubernetes.io/projected/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-kube-api-access-lxgjg\") pod \"redhat-marketplace-qxsj5\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:14 crc kubenswrapper[4862]: I1128 11:58:14.663698 4862 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:15 crc kubenswrapper[4862]: I1128 11:58:15.132521 4862 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxsj5"] Nov 28 11:58:15 crc kubenswrapper[4862]: W1128 11:58:15.141209 4862 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3843477_d5db_4c4b_a75a_ea7c2848bf3d.slice/crio-09f4d9d2e335854362fe8d21ebafbe659e1a5907fd408cfbfe30a30a9f9fddbe WatchSource:0}: Error finding container 09f4d9d2e335854362fe8d21ebafbe659e1a5907fd408cfbfe30a30a9f9fddbe: Status 404 returned error can't find the container with id 09f4d9d2e335854362fe8d21ebafbe659e1a5907fd408cfbfe30a30a9f9fddbe Nov 28 11:58:15 crc kubenswrapper[4862]: I1128 11:58:15.747801 4862 generic.go:334] "Generic (PLEG): container finished" podID="e3843477-d5db-4c4b-a75a-ea7c2848bf3d" containerID="211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2" exitCode=0 Nov 28 11:58:15 crc kubenswrapper[4862]: I1128 11:58:15.747913 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxsj5" event={"ID":"e3843477-d5db-4c4b-a75a-ea7c2848bf3d","Type":"ContainerDied","Data":"211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2"} Nov 28 11:58:15 crc kubenswrapper[4862]: I1128 11:58:15.748233 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxsj5" event={"ID":"e3843477-d5db-4c4b-a75a-ea7c2848bf3d","Type":"ContainerStarted","Data":"09f4d9d2e335854362fe8d21ebafbe659e1a5907fd408cfbfe30a30a9f9fddbe"} Nov 28 11:58:15 crc kubenswrapper[4862]: I1128 11:58:15.751048 4862 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 11:58:16 crc kubenswrapper[4862]: I1128 11:58:16.761440 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxsj5" event={"ID":"e3843477-d5db-4c4b-a75a-ea7c2848bf3d","Type":"ContainerStarted","Data":"909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8"} Nov 28 11:58:17 crc kubenswrapper[4862]: I1128 11:58:17.776600 4862 generic.go:334] "Generic (PLEG): container finished" podID="e3843477-d5db-4c4b-a75a-ea7c2848bf3d" containerID="909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8" exitCode=0 Nov 28 11:58:17 crc kubenswrapper[4862]: I1128 11:58:17.776642 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxsj5" event={"ID":"e3843477-d5db-4c4b-a75a-ea7c2848bf3d","Type":"ContainerDied","Data":"909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8"} Nov 28 11:58:18 crc kubenswrapper[4862]: E1128 11:58:18.129848 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice/crio-8cb9798ee9d78288e7a9e83452469143997c037786e1d50cac0284c1afef9302\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice\": RecentStats: unable to find data in memory cache]" Nov 28 11:58:18 crc kubenswrapper[4862]: I1128 11:58:18.787928 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxsj5" event={"ID":"e3843477-d5db-4c4b-a75a-ea7c2848bf3d","Type":"ContainerStarted","Data":"9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030"} Nov 28 11:58:18 crc kubenswrapper[4862]: I1128 11:58:18.815667 4862 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qxsj5" podStartSLOduration=2.248082167 podStartE2EDuration="4.815649273s" podCreationTimestamp="2025-11-28 11:58:14 +0000 UTC" firstStartedPulling="2025-11-28 11:58:15.750280669 +0000 UTC m=+2173.282794630" lastFinishedPulling="2025-11-28 11:58:18.317847765 +0000 UTC m=+2175.850361736" observedRunningTime="2025-11-28 11:58:18.81236706 +0000 UTC m=+2176.344881021" watchObservedRunningTime="2025-11-28 11:58:18.815649273 +0000 UTC m=+2176.348163204" Nov 28 11:58:24 crc kubenswrapper[4862]: I1128 11:58:24.664196 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:24 crc kubenswrapper[4862]: I1128 11:58:24.664819 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:24 crc kubenswrapper[4862]: I1128 11:58:24.743944 4862 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:24 crc kubenswrapper[4862]: I1128 11:58:24.925650 4862 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:24 crc kubenswrapper[4862]: I1128 11:58:24.992942 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxsj5"] Nov 28 11:58:26 crc kubenswrapper[4862]: I1128 11:58:26.862535 4862 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qxsj5" podUID="e3843477-d5db-4c4b-a75a-ea7c2848bf3d" containerName="registry-server" containerID="cri-o://9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030" gracePeriod=2 Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.805363 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.872189 4862 generic.go:334] "Generic (PLEG): container finished" podID="e3843477-d5db-4c4b-a75a-ea7c2848bf3d" containerID="9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030" exitCode=0 Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.872236 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxsj5" event={"ID":"e3843477-d5db-4c4b-a75a-ea7c2848bf3d","Type":"ContainerDied","Data":"9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030"} Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.872259 4862 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxsj5" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.872281 4862 scope.go:117] "RemoveContainer" containerID="9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.872267 4862 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxsj5" event={"ID":"e3843477-d5db-4c4b-a75a-ea7c2848bf3d","Type":"ContainerDied","Data":"09f4d9d2e335854362fe8d21ebafbe659e1a5907fd408cfbfe30a30a9f9fddbe"} Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.889067 4862 scope.go:117] "RemoveContainer" containerID="909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.903767 4862 scope.go:117] "RemoveContainer" containerID="211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.927477 4862 scope.go:117] "RemoveContainer" containerID="9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030" Nov 28 11:58:27 crc kubenswrapper[4862]: E1128 11:58:27.927945 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030\": container with ID starting with 9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030 not found: ID does not exist" containerID="9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.928002 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030"} err="failed to get container status \"9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030\": rpc error: code = NotFound desc = could not find container \"9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030\": container with ID starting with 9fbd6409493925613e8b0b0229fcb9a8712b21e543225a9f8be43fdc186ae030 not found: ID does not exist" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.928033 4862 scope.go:117] "RemoveContainer" containerID="909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8" Nov 28 11:58:27 crc kubenswrapper[4862]: E1128 11:58:27.929292 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8\": container with ID starting with 909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8 not found: ID does not exist" containerID="909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.929331 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8"} err="failed to get container status \"909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8\": rpc error: code = NotFound desc = could not find container \"909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8\": container with ID starting with 909fbe6435c05420c323309bd87c31fc93c05c94da0261902daec8040982fcd8 not found: ID does not exist" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.929360 4862 scope.go:117] "RemoveContainer" containerID="211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.930389 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-utilities" (OuterVolumeSpecName: "utilities") pod "e3843477-d5db-4c4b-a75a-ea7c2848bf3d" (UID: "e3843477-d5db-4c4b-a75a-ea7c2848bf3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.928724 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-utilities\") pod \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " Nov 28 11:58:27 crc kubenswrapper[4862]: E1128 11:58:27.930551 4862 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2\": container with ID starting with 211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2 not found: ID does not exist" containerID="211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.930573 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxgjg\" (UniqueName: \"kubernetes.io/projected/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-kube-api-access-lxgjg\") pod \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.930575 4862 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2"} err="failed to get container status \"211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2\": rpc error: code = NotFound desc = could not find container \"211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2\": container with ID starting with 211dbe4f47b50724c3ea7b273d5354f026d0d94ed8e187ac5827ee9de5d9f8c2 not found: ID does not exist" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.930623 4862 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-catalog-content\") pod \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\" (UID: \"e3843477-d5db-4c4b-a75a-ea7c2848bf3d\") " Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.931732 4862 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.936800 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-kube-api-access-lxgjg" (OuterVolumeSpecName: "kube-api-access-lxgjg") pod "e3843477-d5db-4c4b-a75a-ea7c2848bf3d" (UID: "e3843477-d5db-4c4b-a75a-ea7c2848bf3d"). InnerVolumeSpecName "kube-api-access-lxgjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:58:27 crc kubenswrapper[4862]: I1128 11:58:27.951672 4862 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3843477-d5db-4c4b-a75a-ea7c2848bf3d" (UID: "e3843477-d5db-4c4b-a75a-ea7c2848bf3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:58:28 crc kubenswrapper[4862]: I1128 11:58:28.032679 4862 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxgjg\" (UniqueName: \"kubernetes.io/projected/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-kube-api-access-lxgjg\") on node \"crc\" DevicePath \"\"" Nov 28 11:58:28 crc kubenswrapper[4862]: I1128 11:58:28.032711 4862 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3843477-d5db-4c4b-a75a-ea7c2848bf3d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:58:28 crc kubenswrapper[4862]: I1128 11:58:28.208506 4862 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxsj5"] Nov 28 11:58:28 crc kubenswrapper[4862]: I1128 11:58:28.215069 4862 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxsj5"] Nov 28 11:58:28 crc kubenswrapper[4862]: E1128 11:58:28.345803 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice/crio-8cb9798ee9d78288e7a9e83452469143997c037786e1d50cac0284c1afef9302\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3843477_d5db_4c4b_a75a_ea7c2848bf3d.slice\": RecentStats: unable to find data in memory cache]" Nov 28 11:58:28 crc kubenswrapper[4862]: I1128 11:58:28.856629 4862 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3843477-d5db-4c4b-a75a-ea7c2848bf3d" path="/var/lib/kubelet/pods/e3843477-d5db-4c4b-a75a-ea7c2848bf3d/volumes" Nov 28 11:58:38 crc kubenswrapper[4862]: E1128 11:58:38.553979 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice/crio-8cb9798ee9d78288e7a9e83452469143997c037786e1d50cac0284c1afef9302\": RecentStats: unable to find data in memory cache]" Nov 28 11:58:48 crc kubenswrapper[4862]: E1128 11:58:48.775857 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice/crio-8cb9798ee9d78288e7a9e83452469143997c037786e1d50cac0284c1afef9302\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice\": RecentStats: unable to find data in memory cache]" Nov 28 11:58:59 crc kubenswrapper[4862]: E1128 11:58:59.040793 4862 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice/crio-8cb9798ee9d78288e7a9e83452469143997c037786e1d50cac0284c1afef9302\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07acf715_5134_4cac_a8d9_289eeb627ccd.slice\": RecentStats: unable to find data in memory cache]" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112307252024443 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112307253017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112302370016477 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112302370015447 5ustar corecore